 PyTorch Release Notesexperience. In the container, see /workspace/README.md for information about customizing your PyTorch image. For more information about PyTorch, including tutorials, documentation, and examples, see: ‣ PyTorch for NGC containers, when you run a container, the following occurs: ‣ The Docker engine loads the image into a container which runs the software. ‣ You define the runtime resources of the container by in your system depends on the DGX OS version that you installed (for DGX systems), the NGC Cloud Image that was provided by a Cloud Service Provider, or the software that you installed to prepare to run0 码力 | 365 页 | 2.94 MB | 1 年前3 PyTorch Release Notesexperience. In the container, see /workspace/README.md for information about customizing your PyTorch image. For more information about PyTorch, including tutorials, documentation, and examples, see: ‣ PyTorch for NGC containers, when you run a container, the following occurs: ‣ The Docker engine loads the image into a container which runs the software. ‣ You define the runtime resources of the container by in your system depends on the DGX OS version that you installed (for DGX systems), the NGC Cloud Image that was provided by a Cloud Service Provider, or the software that you installed to prepare to run0 码力 | 365 页 | 2.94 MB | 1 年前3
 动手学深度学习 v2.0(continues on next page) 目录 5 (continued from previous page) import torchvision from PIL import Image from torch import nn from torch.nn import functional as F from torch.utils import data from torchvision ImageNet数据集发布,并发起ImageNet挑战赛:要求研究人员从100万个样本中训练模型,以区分1000个不同 类别的对象。ImageNet数据集由斯坦福教授李飞飞小组的研究人员开发,利用谷歌图像搜索(Google Image Search)对每一类图像进行预筛选,并利用亚马逊众包(Amazon Mechanical Turk)来标注每张图片的相关 类别。这种规模是前所未有的。这项被称为ImageNet的挑战赛推动了计算机视觉和机器学习研究的发展,挑 d2l 13.1.1 常用的图像增广方法 在对常用图像增广方法的探索时,我们将使用下面这个尺寸为400 × 500的图像作为示例。 d2l.set_figsize() img = d2l.Image.open('../img/cat1.jpg') d2l.plt.imshow(img); 大多数图像增广方法都具有一定的随机性。为了便于观察图像增广的效果,我们下面定义辅助函数apply。此0 码力 | 797 页 | 29.45 MB | 1 年前3 动手学深度学习 v2.0(continues on next page) 目录 5 (continued from previous page) import torchvision from PIL import Image from torch import nn from torch.nn import functional as F from torch.utils import data from torchvision ImageNet数据集发布,并发起ImageNet挑战赛:要求研究人员从100万个样本中训练模型,以区分1000个不同 类别的对象。ImageNet数据集由斯坦福教授李飞飞小组的研究人员开发,利用谷歌图像搜索(Google Image Search)对每一类图像进行预筛选,并利用亚马逊众包(Amazon Mechanical Turk)来标注每张图片的相关 类别。这种规模是前所未有的。这项被称为ImageNet的挑战赛推动了计算机视觉和机器学习研究的发展,挑 d2l 13.1.1 常用的图像增广方法 在对常用图像增广方法的探索时,我们将使用下面这个尺寸为400 × 500的图像作为示例。 d2l.set_figsize() img = d2l.Image.open('../img/cat1.jpg') d2l.plt.imshow(img); 大多数图像增广方法都具有一定的随机性。为了便于观察图像增广的效果,我们下面定义辅助函数apply。此0 码力 | 797 页 | 29.45 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 3 - Learning Techniquestarget label is a composite of the inputs that were combined. A combination of a dog with a hamster image (figure 3-5) is assigned a composite [dog, hamster] label! 2 A whale’s tail fins are called flukes problems. Figure 3-5: A mixed composite of a dog (30%) and a hamster (70%). The label assigned to this image is a composite of the two classes in the same proportion. Thus, the model would be expected to predict a dataset Nx the size? What are the constraining factors? An image transformation recomputes the pixel values. The rotation of an RGB image of 100x100 requires at least 100x100x3 (3 channels) computations0 码力 | 56 页 | 18.93 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 3 - Learning Techniquestarget label is a composite of the inputs that were combined. A combination of a dog with a hamster image (figure 3-5) is assigned a composite [dog, hamster] label! 2 A whale’s tail fins are called flukes problems. Figure 3-5: A mixed composite of a dog (30%) and a hamster (70%). The label assigned to this image is a composite of the two classes in the same proportion. Thus, the model would be expected to predict a dataset Nx the size? What are the constraining factors? An image transformation recomputes the pixel values. The rotation of an RGB image of 100x100 requires at least 100x100x3 (3 channels) computations0 码力 | 56 页 | 18.93 MB | 1 年前3
 keras tutorialin data science fields like robotics, artificial intelligence(AI), audio & video recognition and image recognition. Artificial neural network is the core of deep learning methodologies. Deep learning keras/keras.json. keras.json { "image_data_format": "channels_last", "epsilon": 1e-07, "floatx": "float32", "backend": "tensorflow" } Here,  image_data_format represent the data format just change the backend = theano in keras.json file. It is described below: keras.json { "image_data_format": "channels_last", "epsilon": 1e-07, "floatx": "float32", "backend": "theano"0 码力 | 98 页 | 1.57 MB | 1 年前3 keras tutorialin data science fields like robotics, artificial intelligence(AI), audio & video recognition and image recognition. Artificial neural network is the core of deep learning methodologies. Deep learning keras/keras.json. keras.json { "image_data_format": "channels_last", "epsilon": 1e-07, "floatx": "float32", "backend": "tensorflow" } Here,  image_data_format represent the data format just change the backend = theano in keras.json file. It is described below: keras.json { "image_data_format": "channels_last", "epsilon": 1e-07, "floatx": "float32", "backend": "theano"0 码力 | 98 页 | 1.57 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniqueshigh quality image of a cat. The cat on the right is a lower quality compressed image. Source Both the cat images in figure 2-2 might serve their purpose equally well, but the compressed image is an order smaller. Discrete Cosine Transform (DCT), is a popular algorithm which is used in the JPEG format for image compression, and the MP3 format for audio. DCT breaks down the given input data into independent components transmitting images back to earth. However, transmission costs make it infeasible to send the original image. Can we compress the transmission, and decompress it on arrival? If so, what would be the ideal tradeoff0 码力 | 33 页 | 1.96 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniqueshigh quality image of a cat. The cat on the right is a lower quality compressed image. Source Both the cat images in figure 2-2 might serve their purpose equally well, but the compressed image is an order smaller. Discrete Cosine Transform (DCT), is a popular algorithm which is used in the JPEG format for image compression, and the MP3 format for audio. DCT breaks down the given input data into independent components transmitting images back to earth. However, transmission costs make it infeasible to send the original image. Can we compress the transmission, and decompress it on arrival? If so, what would be the ideal tradeoff0 码力 | 33 页 | 1.96 MB | 1 年前3
 Experiment 6: K-MeansK-Means November 27, 2018 1 Description In this exercise, you will use K-means to compress an image by reducing the number of colors it contains. To begin, download data6.zip and unpack its contents to Frank Wouters and is used with his permission. 2 Image Representation The data pack for this exercise contains a 538-pixel by 538-pixel TIFF image named bird large.tiff. It looks like the picture below below. In a straightforward 24-bit color representation of this image, each pixel is represented as three 8-bit numbers (ranging from 0 to 255) that specify red, green and blue intensity values. Our bird0 码力 | 3 页 | 605.46 KB | 1 年前3 Experiment 6: K-MeansK-Means November 27, 2018 1 Description In this exercise, you will use K-means to compress an image by reducing the number of colors it contains. To begin, download data6.zip and unpack its contents to Frank Wouters and is used with his permission. 2 Image Representation The data pack for this exercise contains a 538-pixel by 538-pixel TIFF image named bird large.tiff. It looks like the picture below below. In a straightforward 24-bit color representation of this image, each pixel is represented as three 8-bit numbers (ranging from 0 to 255) that specify red, green and blue intensity values. Our bird0 码力 | 3 页 | 605.46 KB | 1 年前3
 Keras: 基于 Python 的深度学习库tower_3], axis=1) 3.2.7.2 卷积层上的残差连接 有关残差网络 (Residual Network) 的更多信息,请参阅 Deep Residual Learning for Image Recogni- tion。 from keras.layers import Conv2D, Input # 输入张量为 3 通道 256x256 图像 x = Input(shape=(256 add(MaxPooling2D((2, 2))) vision_model.add(Flatten()) # 现在让我们用视觉模型来得到一个输出张量: image_input = Input(shape=(224, 224, 3)) encoded_image = vision_model(image_input) # 接下来,定义一个语言模型来将问题编码成一个向量。 # 每个问题最长 100 个词,词的索引从 1 到 concatenate([encoded_question, encoded_image]) # 然后在上面训练一个 1000 词的逻辑回归模型: output = Dense(1000, activation='softmax')(merged) # 最终模型: vqa_model = Model(inputs=[image_input, question_input], outputs=output)0 码力 | 257 页 | 1.19 MB | 1 年前3 Keras: 基于 Python 的深度学习库tower_3], axis=1) 3.2.7.2 卷积层上的残差连接 有关残差网络 (Residual Network) 的更多信息,请参阅 Deep Residual Learning for Image Recogni- tion。 from keras.layers import Conv2D, Input # 输入张量为 3 通道 256x256 图像 x = Input(shape=(256 add(MaxPooling2D((2, 2))) vision_model.add(Flatten()) # 现在让我们用视觉模型来得到一个输出张量: image_input = Input(shape=(224, 224, 3)) encoded_image = vision_model(image_input) # 接下来,定义一个语言模型来将问题编码成一个向量。 # 每个问题最长 100 个词,词的索引从 1 到 concatenate([encoded_question, encoded_image]) # 然后在上面训练一个 1000 词的逻辑回归模型: output = Dense(1000, activation='softmax')(merged) # 最终模型: vqa_model = Model(inputs=[image_input, question_input], outputs=output)0 码力 | 257 页 | 1.19 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 7 - Automationand channel configurations can also be parameterized using hyperparameters. For example, when using image data augmentation with rotation, we can treat the angle of rotation as a hyper-parameter. Think of chapter 3. # Dataset image size IMG_SIZE = 264 def resize_image(image, label): image = tf.image.resize(image, [IMG_SIZE, IMG_SIZE]) image = tf.cast(image, tf.uint8) return image, label train_ds = train_ds train_ds.map(resize_image) val_ds = val_ds.map(resize_image) test_ds = test_ds.map(resize_image) Note that the create_model() function here has two additional parameters: learning_rate and dropout_rate0 码力 | 33 页 | 2.48 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 7 - Automationand channel configurations can also be parameterized using hyperparameters. For example, when using image data augmentation with rotation, we can treat the angle of rotation as a hyper-parameter. Think of chapter 3. # Dataset image size IMG_SIZE = 264 def resize_image(image, label): image = tf.image.resize(image, [IMG_SIZE, IMG_SIZE]) image = tf.cast(image, tf.uint8) return image, label train_ds = train_ds train_ds.map(resize_image) val_ds = val_ds.map(resize_image) test_ds = test_ds.map(resize_image) Note that the create_model() function here has two additional parameters: learning_rate and dropout_rate0 码力 | 33 页 | 2.48 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architectureshumans can intuitively grasp similarities between different objects. For instance, when we see an image of a dog or a cat, it is likely that we would find them both to be cute. However, a picture of a snake following goals: a) To compress the information content of high-dimensional concepts such as text, image, audio, video, etc. to a low-dimensional representation such as a fixed length vector of floating perspective of training the model, it is agnostic to what the embedding is for (a piece of text, audio, image, video, or some abstract concept). Here is a quick recipe to train embedding-based models: 1. Embedding0 码力 | 53 页 | 3.92 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architectureshumans can intuitively grasp similarities between different objects. For instance, when we see an image of a dog or a cat, it is likely that we would find them both to be cute. However, a picture of a snake following goals: a) To compress the information content of high-dimensional concepts such as text, image, audio, video, etc. to a low-dimensional representation such as a fixed length vector of floating perspective of training the model, it is agnostic to what the embedding is for (a piece of text, audio, image, video, or some abstract concept). Here is a quick recipe to train embedding-based models: 1. Embedding0 码力 | 53 页 | 3.92 MB | 1 年前3
 Lecture 5: Gaussian Discriminant Analysis, Naive Bayesand EM September 27, 2023 25 / 122 Warm Up (Contd.) Task: Identify if there is a cat in a given image. Feng Li (SDU) GDA, NB and EM September 27, 2023 26 / 122 Warm Up (Contd.) Images (to be classified random An image is represented by a vector of features The feature vectors are random, since the images are randomly given Random variable X representing the feature vector (and thus the image) The labels = x): Given an image X = x (whose feature is x), what is the probability of Y = y (with y = 1 denoting there is a cat and y = 0 denoting there is not)? P(X = x | Y = y): Given an image with Y = y (whose0 码力 | 122 页 | 1.35 MB | 1 年前3 Lecture 5: Gaussian Discriminant Analysis, Naive Bayesand EM September 27, 2023 25 / 122 Warm Up (Contd.) Task: Identify if there is a cat in a given image. Feng Li (SDU) GDA, NB and EM September 27, 2023 26 / 122 Warm Up (Contd.) Images (to be classified random An image is represented by a vector of features The feature vectors are random, since the images are randomly given Random variable X representing the feature vector (and thus the image) The labels = x): Given an image X = x (whose feature is x), what is the probability of Y = y (with y = 1 denoting there is a cat and y = 0 denoting there is not)? P(X = x | Y = y): Given an image with Y = y (whose0 码力 | 122 页 | 1.35 MB | 1 年前3
共 40 条
- 1
- 2
- 3
- 4













