 PyTorch Release Notesdebugging experience. In the container, see /workspace/README.md for information about customizing your PyTorch image. For more information about PyTorch, including tutorials, documentation, and examples, see: ‣ PyTorch website ‣ PyTorch project This document provides information about the key features, software enhancements and improvements, known issues, and how to run this container. PyTorch RN-08516-001_v23 access and can log in to the NGC container registry. Refer to NGC Getting Started Guide for more information. The deep learning frameworks, the NGC Docker containers, and the deep learning framework containers0 码力 | 365 页 | 2.94 MB | 1 年前3 PyTorch Release Notesdebugging experience. In the container, see /workspace/README.md for information about customizing your PyTorch image. For more information about PyTorch, including tutorials, documentation, and examples, see: ‣ PyTorch website ‣ PyTorch project This document provides information about the key features, software enhancements and improvements, known issues, and how to run this container. PyTorch RN-08516-001_v23 access and can log in to the NGC container registry. Refer to NGC Getting Started Guide for more information. The deep learning frameworks, the NGC Docker containers, and the deep learning framework containers0 码力 | 365 页 | 2.94 MB | 1 年前3
 AI大模型千问 qwen 中文文档install vllm 运行以下代码以构建 vllm 服务。此处我们以 Qwen1.5-7B-Chat 为例: python -m vllm.entrypoints.openai.api_server --model Qwen/Qwen1.5-7B-Chat 然后,您可以使用 create chat interface 来与 Qwen 进行交流: curl http://localhos 包中的 Python 客户端: from openai import OpenAI # Set OpenAI's API key and API base to use vLLM's API server. openai_api_key = "EMPTY" openai_api_base = "http://localhost:8000/v1" client = OpenAI( (续下页) creating and quantizing GGUF files in quantization/llama.cpp. You can refer to that document for more information. 1.4.4 PPL 评测 llama.cpp 为我们提供了评估 GGUF 模型 PPL 性能的方法。为了实现这一点,你需要准备一个数据集,比如 “wiki 测试”。这里我们展示了一个运行测试的例子。0 码力 | 56 页 | 835.78 KB | 1 年前3 AI大模型千问 qwen 中文文档install vllm 运行以下代码以构建 vllm 服务。此处我们以 Qwen1.5-7B-Chat 为例: python -m vllm.entrypoints.openai.api_server --model Qwen/Qwen1.5-7B-Chat 然后,您可以使用 create chat interface 来与 Qwen 进行交流: curl http://localhos 包中的 Python 客户端: from openai import OpenAI # Set OpenAI's API key and API base to use vLLM's API server. openai_api_key = "EMPTY" openai_api_base = "http://localhost:8000/v1" client = OpenAI( (续下页) creating and quantizing GGUF files in quantization/llama.cpp. You can refer to that document for more information. 1.4.4 PPL 评测 llama.cpp 为我们提供了评估 GGUF 模型 PPL 性能的方法。为了实现这一点,你需要准备一个数据集,比如 “wiki 测试”。这里我们展示了一个运行测试的例子。0 码力 | 56 页 | 835.78 KB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 1 - IntroductionHinton. "Imagenet classification with deep convolutional neural networks." Advances in neural information processing systems 25 (2012): 1097-1105. do linear algebra operations such as multiplying two Having such a toolbox to make our models pareto-optimal has the following benefits: Sustainable Server-Side Scaling Training and deploying large deep learning models is costly. While training is a one-time inference can be run completely on the user’s device without the need to send the input data to the server-side. New Applications Efficiency would also enable applications that couldn’t have otherwise been0 码力 | 21 页 | 3.17 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 1 - IntroductionHinton. "Imagenet classification with deep convolutional neural networks." Advances in neural information processing systems 25 (2012): 1097-1105. do linear algebra operations such as multiplying two Having such a toolbox to make our models pareto-optimal has the following benefits: Sustainable Server-Side Scaling Training and deploying large deep learning models is costly. While training is a one-time inference can be run completely on the user’s device without the need to send the input data to the server-side. New Applications Efficiency would also enable applications that couldn’t have otherwise been0 码力 | 21 页 | 3.17 MB | 1 年前3
 keras tutorial[MSC v.1900 64 bit (AMD64)] on win32 Type "help", "copyright", "credits" or "license" for more information. >>> As of now the latest version is ‘3.7.2’. If Python is not installed, then visit the official analyzing the input in layer by layer manner, where each layer progressively extracts higher level information about the input. Let us take a simple scenario of analyzing an image. Let us assume that your main role of axon is to transmit information from one neuron to another to which it is connected. Similarly, the main role of dendrites is to receive the information being transmitted by the axons of0 码力 | 98 页 | 1.57 MB | 1 年前3 keras tutorial[MSC v.1900 64 bit (AMD64)] on win32 Type "help", "copyright", "credits" or "license" for more information. >>> As of now the latest version is ‘3.7.2’. If Python is not installed, then visit the official analyzing the input in layer by layer manner, where each layer progressively extracts higher level information about the input. Let us take a simple scenario of analyzing an image. Let us assume that your main role of axon is to transmit information from one neuron to another to which it is connected. Similarly, the main role of dendrites is to receive the information being transmitted by the axons of0 码力 | 98 页 | 1.57 MB | 1 年前3
 【PyTorch深度学习-龙龙老师】-测试版202112LarochelleH., BeygelzimerA., d\textquotesingle Alch é-BucF., FoxE., & GarnettR. (编辑), Advances in Neural Information Processing Systems 32 (页 8024–8035). Curran Associates, Inc. 检索来源: http://papers.neurips.cc/ Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” 出处 Advances in Neural Information Processing Systems 25, F. Pereira, C. J. C. Burges, L. Bottou 和 K. Q. Weinberger, 编辑, Curran Warde-Farley, S. Ozair, A. Courville 和 Y. Bengio, “Generative Adversarial Nets,” 出处 Advances in Neural Information Processing Systems 27, Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence 和 K. Q. Weinberger0 码力 | 439 页 | 29.91 MB | 1 年前3 【PyTorch深度学习-龙龙老师】-测试版202112LarochelleH., BeygelzimerA., d\textquotesingle Alch é-BucF., FoxE., & GarnettR. (编辑), Advances in Neural Information Processing Systems 32 (页 8024–8035). Curran Associates, Inc. 检索来源: http://papers.neurips.cc/ Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” 出处 Advances in Neural Information Processing Systems 25, F. Pereira, C. J. C. Burges, L. Bottou 和 K. Q. Weinberger, 编辑, Curran Warde-Farley, S. Ozair, A. Courville 和 Y. Bengio, “Generative Adversarial Nets,” 出处 Advances in Neural Information Processing Systems 27, Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence 和 K. Q. Weinberger0 码力 | 439 页 | 29.91 MB | 1 年前3
 动手学深度学习 v2.0loss),它是分类问题最常用的损失之一。本节我们将通过介绍信息论基础来理解交叉熵损失。如果 想了解更多信息论的细节,请进一步参考 本书附录中关于信息论的一节52。 3.4.7 信息论基础 信息论(information theory)涉及编码、解码、发送以及尽可能简洁地处理信息或数据。 熵 信息论的核心思想是量化数据中的信息内容。在信息论中,该数值被称为分布P的熵(entropy)。可以通过 以下方程得到: learning/distributions.html 52 https://d2l.ai/chapter_appendix‐mathematics‐for‐deep‐learning/information‐theory.html 3.4. softmax回归 109 信息量 压缩与预测有什么关系呢?想象一下,我们有一个要压缩的数据流。如果我们很容易预测下一个数据,那么 这个数据就很容 型。 • 我们可以使用困惑度来评价语言模型的质量。 103 https://d2l.ai/chapter_appendix‐mathematics‐for‐deep‐learning/information‐theory.html 316 8. 循环神经网络 练习 1. 如果我们使用循环神经网络来预测文本序列中的下一个字符,那么任意输出所需的维度是多少? 2. 为什么循环神经网络可以基0 码力 | 797 页 | 29.45 MB | 1 年前3 动手学深度学习 v2.0loss),它是分类问题最常用的损失之一。本节我们将通过介绍信息论基础来理解交叉熵损失。如果 想了解更多信息论的细节,请进一步参考 本书附录中关于信息论的一节52。 3.4.7 信息论基础 信息论(information theory)涉及编码、解码、发送以及尽可能简洁地处理信息或数据。 熵 信息论的核心思想是量化数据中的信息内容。在信息论中,该数值被称为分布P的熵(entropy)。可以通过 以下方程得到: learning/distributions.html 52 https://d2l.ai/chapter_appendix‐mathematics‐for‐deep‐learning/information‐theory.html 3.4. softmax回归 109 信息量 压缩与预测有什么关系呢?想象一下,我们有一个要压缩的数据流。如果我们很容易预测下一个数据,那么 这个数据就很容 型。 • 我们可以使用困惑度来评价语言模型的质量。 103 https://d2l.ai/chapter_appendix‐mathematics‐for‐deep‐learning/information‐theory.html 316 8. 循环神经网络 练习 1. 如果我们使用循环神经网络来预测文本序列中的下一个字符,那么任意输出所需的维度是多少? 2. 为什么循环神经网络可以基0 码力 | 797 页 | 29.45 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architecturesrepresent aspects of an input numerically. It must fulfill the following goals: a) To compress the information content of high-dimensional concepts such as text, image, audio, video, etc. to a low-dimensional animals occupy the top-left area of the plot. Note how we have compressed the high-dimensional information about animals into just two dimensions, and established a relationship between them purely using after having read this in many textbooks throughout our life. Hopefully, we have given enough information to make this actually straightforward. as the number of features that our model learns for each0 码力 | 53 页 | 3.92 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architecturesrepresent aspects of an input numerically. It must fulfill the following goals: a) To compress the information content of high-dimensional concepts such as text, image, audio, video, etc. to a low-dimensional animals occupy the top-left area of the plot. Note how we have compressed the high-dimensional information about animals into just two dimensions, and established a relationship between them purely using after having read this in many textbooks throughout our life. Hopefully, we have given enough information to make this actually straightforward. as the number of features that our model learns for each0 码力 | 53 页 | 3.92 MB | 1 年前3
 Lecture 1: OverviewMobile Computing, Internet of Things. Feng Li (SDU) Overview September 6, 2023 3 / 57 Course Information We will investigate fundamental concepts, techniques and algorithms in machine learning. The topics September 6, 2023 5 / 57 Prerequisite Courses Linear algebra Calculus Probability and Statistics Information theory Convex Optimization Feng Li (SDU) Overview September 6, 2023 6 / 57 Remarks Lectures knowledge from large databases (data mining) Market basket analysis (e.g. diapers and beer) Medical information mining (e.g. migraines to calcium channel blockers to magnesium) Computational studies of learning0 码力 | 57 页 | 2.41 MB | 1 年前3 Lecture 1: OverviewMobile Computing, Internet of Things. Feng Li (SDU) Overview September 6, 2023 3 / 57 Course Information We will investigate fundamental concepts, techniques and algorithms in machine learning. The topics September 6, 2023 5 / 57 Prerequisite Courses Linear algebra Calculus Probability and Statistics Information theory Convex Optimization Feng Li (SDU) Overview September 6, 2023 6 / 57 Remarks Lectures knowledge from large databases (data mining) Market basket analysis (e.g. diapers and beer) Medical information mining (e.g. migraines to calcium channel blockers to magnesium) Computational studies of learning0 码力 | 57 页 | 2.41 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 3 - Learning TechniquesSuch a shift has two side-effects. First, a part of the image “falls off” the top edge. That information will be lost. And the second, the lower part of the image doesn’t have any pixel data because it the press of his successful flights. The author has taken grammatical liberties to compress the information to save dollars. In spite of the obvious errors, the message is loud and clear. Figure 3-9: This an elementary level Spanish speaker’s response, “estoy ir mercado”, sufficiently conveys the information that the person is going to the market. A version of this example could be a native english speaker’s0 码力 | 56 页 | 18.93 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 3 - Learning TechniquesSuch a shift has two side-effects. First, a part of the image “falls off” the top edge. That information will be lost. And the second, the lower part of the image doesn’t have any pixel data because it the press of his successful flights. The author has taken grammatical liberties to compress the information to save dollars. In spite of the obvious errors, the message is loud and clear. Figure 3-9: This an elementary level Spanish speaker’s response, “estoy ir mercado”, sufficiently conveys the information that the person is going to the market. A version of this example could be a native english speaker’s0 码力 | 56 页 | 18.93 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniquesin the history of computing, scientists have worked tirelessly towards storing and transmitting information in as few bits as possible. Depending on the use case, we might be interested in compressing in are losing some information as a trade off. It is especially applicable for multimedia (audio, video, images) data,, where it is likely that either humans who will consume the information will not notice notice the loss of some information, or do not necessarily care about the loss in quality. Figure 2-2: On the left is a high quality image of a cat. The cat on the right is a lower quality compressed image0 码力 | 33 页 | 1.96 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniquesin the history of computing, scientists have worked tirelessly towards storing and transmitting information in as few bits as possible. Depending on the use case, we might be interested in compressing in are losing some information as a trade off. It is especially applicable for multimedia (audio, video, images) data,, where it is likely that either humans who will consume the information will not notice notice the loss of some information, or do not necessarily care about the loss in quality. Figure 2-2: On the left is a high quality image of a cat. The cat on the right is a lower quality compressed image0 码力 | 33 页 | 1.96 MB | 1 年前3
共 37 条
- 1
- 2
- 3
- 4













