 《Efficient Deep Learning Book》[EDL] Chapter 1 - IntroductionHaving such a toolbox to make our models pareto-optimal has the following benefits: Sustainable Server-Side Scaling Training and deploying large deep learning models is costly. While training is a one-time inference can be run completely on the user’s device without the need to send the input data to the server-side. New Applications Efficiency would also enable applications that couldn’t have otherwise been device. This further reduces the available resources for a single model. This could happen on the server-side where multiple models are co-located on the same machine, or could be in an app where different0 码力 | 21 页 | 3.17 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 1 - IntroductionHaving such a toolbox to make our models pareto-optimal has the following benefits: Sustainable Server-Side Scaling Training and deploying large deep learning models is costly. While training is a one-time inference can be run completely on the user’s device without the need to send the input data to the server-side. New Applications Efficiency would also enable applications that couldn’t have otherwise been device. This further reduces the available resources for a single model. This could happen on the server-side where multiple models are co-located on the same machine, or could be in an app where different0 码力 | 21 页 | 3.17 MB | 1 年前3
 AI大模型千问 qwen 中文文档install vllm 运行以下代码以构建 vllm 服务。此处我们以 Qwen1.5-7B-Chat 为例: python -m vllm.entrypoints.openai.api_server --model Qwen/Qwen1.5-7B-Chat 然后,您可以使用 create chat interface 来与 Qwen 进行交流: curl http://localhos 包中的 Python 客户端: from openai import OpenAI # Set OpenAI's API key and API base to use vLLM's API server. openai_api_key = "EMPTY" openai_api_base = "http://localhost:8000/v1" client = OpenAI( (续下页) safetensor.index.json │ │ ├── merges.txt │ │ ├── tokenizer_config.json │ │ └── vocab.json 随后你需要运行 python server.py 来启动你的网页服务。请点击进入 `http://localhost:7860/?__theme=dark` 然后享受使用 Qwen 的 Web UI 吧! 1.6.2 下一步 TGW0 码力 | 56 页 | 835.78 KB | 1 年前3 AI大模型千问 qwen 中文文档install vllm 运行以下代码以构建 vllm 服务。此处我们以 Qwen1.5-7B-Chat 为例: python -m vllm.entrypoints.openai.api_server --model Qwen/Qwen1.5-7B-Chat 然后,您可以使用 create chat interface 来与 Qwen 进行交流: curl http://localhos 包中的 Python 客户端: from openai import OpenAI # Set OpenAI's API key and API base to use vLLM's API server. openai_api_key = "EMPTY" openai_api_base = "http://localhost:8000/v1" client = OpenAI( (续下页) safetensor.index.json │ │ ├── merges.txt │ │ ├── tokenizer_config.json │ │ └── vocab.json 随后你需要运行 python server.py 来启动你的网页服务。请点击进入 `http://localhost:7860/?__theme=dark` 然后享受使用 Qwen 的 Web UI 吧! 1.6.2 下一步 TGW0 码力 | 56 页 | 835.78 KB | 1 年前3
 PyTorch Release Notesspeed as a deep learning framework and provides accelerated NumPy-like functionality. PyTorch also includes standard defined neural network layers, deep learning optimizers, data loading utilities, and multi-gpu environment (/usr/local/lib/ python3.10/dist-packages/torch) in the container image. The container also includes the following: ‣ Ubuntu 22.04 including Python 3.10 ‣ NVIDIA CUDA® 12.1.1 ‣ NVIDIA cuBLAS 12 see Deep Learning Frameworks Support Matrix. Key Features and Enhancements This PyTorch release includes the following key features and enhancements. ‣ PyTorch container image version 23.07 is based on0 码力 | 365 页 | 2.94 MB | 1 年前3 PyTorch Release Notesspeed as a deep learning framework and provides accelerated NumPy-like functionality. PyTorch also includes standard defined neural network layers, deep learning optimizers, data loading utilities, and multi-gpu environment (/usr/local/lib/ python3.10/dist-packages/torch) in the container image. The container also includes the following: ‣ Ubuntu 22.04 including Python 3.10 ‣ NVIDIA CUDA® 12.1.1 ‣ NVIDIA cuBLAS 12 see Deep Learning Frameworks Support Matrix. Key Features and Enhancements This PyTorch release includes the following key features and enhancements. ‣ PyTorch container image version 23.07 is based on0 码力 | 365 页 | 2.94 MB | 1 年前3
 keras tutorialprocessing, etc.,) and the available input data. Algorithm is represented by Model in Keras. Algorithm includes one or more layers. Each layers in ANN can be represented by Keras Layer in Keras.  Prepare download the online machine learning data for training purposes. It fetches the data from online server, process the data and return the data as training and test set. Let us check the data provided by keras dataset module.  Line 3 calls the load_data function, which will fetch the data from online server and return the data as 2 tuples, First tuple, (x_train, y_train) represent the Keras0 码力 | 98 页 | 1.57 MB | 1 年前3 keras tutorialprocessing, etc.,) and the available input data. Algorithm is represented by Model in Keras. Algorithm includes one or more layers. Each layers in ANN can be represented by Keras Layer in Keras.  Prepare download the online machine learning data for training purposes. It fetches the data from online server, process the data and return the data as training and test set. Let us check the data provided by keras dataset module.  Line 3 calls the load_data function, which will fetch the data from online server and return the data as 2 tuples, First tuple, (x_train, y_train) represent the Keras0 码力 | 98 页 | 1.57 MB | 1 年前3
 PyTorch Tutorialconda install pytorch -c pytorch • Others via pip: pip3 install torch • ???????????? On Princeton CS server (ssh cycles.cs.princeton.edu) • Non-CS students can request a class account. • Miniconda is highly jupyter • ???????????? Run on your computer • jupyter notebook • ???????????? Run on Princeton CS server • Pick any 4-digit number, say 1234 • ???????????? hostname -s • ???????????? jupyter notebook --no-browser dynamic computational graphs. Thus a user can change them during runtime. • It includes many layers as Torch. • It includes lot of loss functions. • It allows building networks whose structure is dependent0 码力 | 38 页 | 4.09 MB | 1 年前3 PyTorch Tutorialconda install pytorch -c pytorch • Others via pip: pip3 install torch • ???????????? On Princeton CS server (ssh cycles.cs.princeton.edu) • Non-CS students can request a class account. • Miniconda is highly jupyter • ???????????? Run on your computer • jupyter notebook • ???????????? Run on Princeton CS server • Pick any 4-digit number, say 1234 • ???????????? hostname -s • ???????????? jupyter notebook --no-browser dynamic computational graphs. Thus a user can change them during runtime. • It includes many layers as Torch. • It includes lot of loss functions. • It allows building networks whose structure is dependent0 码力 | 38 页 | 4.09 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniquesfrom a 32-bit floating point value to a b-bit integer, and back again, we need a mapping from one side to the other. It is easy to learn a mapping from 32-bit to b-bit values. We would also want to keep compiles the model by providing the necessary components the framework needs to train the model. This includes the loss function, the optimizer, and finally the metrics. def get_compiled_model(): """Create0 码力 | 33 页 | 1.96 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 2 - Compression Techniquesfrom a 32-bit floating point value to a b-bit integer, and back again, we need a mapping from one side to the other. It is easy to learn a mapping from 32-bit to b-bit values. We would also want to keep compiles the model by providing the necessary components the framework needs to train the model. This includes the loss function, the optimizer, and finally the metrics. def get_compiled_model(): """Create0 码力 | 33 页 | 1.96 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architecturesthe conditional probability , where the size of the sliding window of context is ( words on each side of the masked word). Let’s take an example with such that our window size is . In this scenario, chapter was focussed on two different sets of architectures. The first set of architectures which includes embeddings and attention leverage learning the data distribution better to improve model quality0 码力 | 53 页 | 3.92 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architecturesthe conditional probability , where the size of the sliding window of context is ( words on each side of the masked word). Let’s take an example with such that our window size is . In this scenario, chapter was focussed on two different sets of architectures. The first set of architectures which includes embeddings and attention leverage learning the data distribution better to improve model quality0 码力 | 53 页 | 3.92 MB | 1 年前3
 阿里云上深度学习建模实践-程孟力标准化模型库  标准化解决方案 1.方案复杂 图像 搜索 推荐 语音 视频理解 NLP 广告 CNN RNN GNN MLP Tensorflow PyTorch Parameter Server MPI TreeModel SQL MapReduce Blink  场景丰富: 图像/视频/推荐/搜索  大数据+大模型: Model Zoo  跨场景+跨模态  开箱即用: 卡证OCR • 人脸检测 • 活体检测 •人脸比对 Mobile SDK API + customer 示例: e-Know Your Customer eKYC eKYC Server eKYC SDK/API  多语言、国际化  多种证件版式  准确率领先同类产品  集成方便 标准化: Standard Solutions 智能推荐解决方案: 推荐请求 PAI-Studio–建模平台 Graph优化 [User Graph去重] 内存Allocate优化 ParallelStringOp [split/type conversion] Sequence Feature [side info] Op Fusion [hash + embedding] Overlap Execution [FG OP化] Item Feature增量更新 3.工程优化复 杂 40 码力 | 40 页 | 8.51 MB | 1 年前3 阿里云上深度学习建模实践-程孟力标准化模型库  标准化解决方案 1.方案复杂 图像 搜索 推荐 语音 视频理解 NLP 广告 CNN RNN GNN MLP Tensorflow PyTorch Parameter Server MPI TreeModel SQL MapReduce Blink  场景丰富: 图像/视频/推荐/搜索  大数据+大模型: Model Zoo  跨场景+跨模态  开箱即用: 卡证OCR • 人脸检测 • 活体检测 •人脸比对 Mobile SDK API + customer 示例: e-Know Your Customer eKYC eKYC Server eKYC SDK/API  多语言、国际化  多种证件版式  准确率领先同类产品  集成方便 标准化: Standard Solutions 智能推荐解决方案: 推荐请求 PAI-Studio–建模平台 Graph优化 [User Graph去重] 内存Allocate优化 ParallelStringOp [split/type conversion] Sequence Feature [side info] Op Fusion [hash + embedding] Overlap Execution [FG OP化] Item Feature增量更新 3.工程优化复 杂 40 码力 | 40 页 | 8.51 MB | 1 年前3
 机器学习课程-温州大学-08机器学习-集成学习4)分布式支持,可快速处理海量数据 37 4.LightGBM LightGBM 的主要改进 LightGBM与XGBoost相比,主要有以下几个改进: • 基于梯度的单边采样算法(Gradient-based One-Side Sampling, GOSS); • 互斥特征捆绑算法(Exclusive Feature Bundling, EFB); • 直方图算法( Histogram ); • 基于最大深度的 Leaf-wise 的垂直生长算法; LightGBM = XGBoost + GOSS + EFB+ Histogram 38 4.LightGBM 基于梯度的单边采样算法(Gradient-based One-Side Sampling, GOSS) 主要思想是通过对样本采样的方法来减少计算目标函数增益时候的复杂度。 GOSS 算法保留了梯度大的样本,并对梯度小的样本进行随机抽样,为了不改 变样本的数据分 如果一个样本的梯度很小,说明该样本的训练误差很小,或者说该样本已经得 到了很好的训练(well-trained)。 39 4.LightGBM 基于梯度的单边采样算法(Gradient-based One-Side Sampling, GOSS) 输入:训练数据,迭代步数d,大梯度数据的采样率a,小梯度数据的采样率b,损失函数和若 学习器的类型(一般为决策树) 输出:训练好的强学习器 (1)根据样本点的梯度的绝对值对它们进行降序排序;0 码力 | 50 页 | 2.03 MB | 1 年前3 机器学习课程-温州大学-08机器学习-集成学习4)分布式支持,可快速处理海量数据 37 4.LightGBM LightGBM 的主要改进 LightGBM与XGBoost相比,主要有以下几个改进: • 基于梯度的单边采样算法(Gradient-based One-Side Sampling, GOSS); • 互斥特征捆绑算法(Exclusive Feature Bundling, EFB); • 直方图算法( Histogram ); • 基于最大深度的 Leaf-wise 的垂直生长算法; LightGBM = XGBoost + GOSS + EFB+ Histogram 38 4.LightGBM 基于梯度的单边采样算法(Gradient-based One-Side Sampling, GOSS) 主要思想是通过对样本采样的方法来减少计算目标函数增益时候的复杂度。 GOSS 算法保留了梯度大的样本,并对梯度小的样本进行随机抽样,为了不改 变样本的数据分 如果一个样本的梯度很小,说明该样本的训练误差很小,或者说该样本已经得 到了很好的训练(well-trained)。 39 4.LightGBM 基于梯度的单边采样算法(Gradient-based One-Side Sampling, GOSS) 输入:训练数据,迭代步数d,大梯度数据的采样率a,小梯度数据的采样率b,损失函数和若 学习器的类型(一般为决策树) 输出:训练好的强学习器 (1)根据样本点的梯度的绝对值对它们进行降序排序;0 码力 | 50 页 | 2.03 MB | 1 年前3
 《Efficient Deep Learning Book》[EDL] Chapter 6 - Advanced Learning Techniques - Technical Reviewlocal minimas. The left hand side image shows a loss landscape that has a sharp and steep minima because the loss suddenly drops to a very low value. On the right hand side shows the loss landscape that its predictions over the same classes as the teacher. We now have a loss function on the student side, where the labels are the subclass logits from the teacher, and the predictions are the subclass0 码力 | 31 页 | 4.03 MB | 1 年前3 《Efficient Deep Learning Book》[EDL] Chapter 6 - Advanced Learning Techniques - Technical Reviewlocal minimas. The left hand side image shows a loss landscape that has a sharp and steep minima because the loss suddenly drops to a very low value. On the right hand side shows the loss landscape that its predictions over the same classes as the teacher. We now have a loss function on the student side, where the labels are the subclass logits from the teacher, and the predictions are the subclass0 码力 | 31 页 | 4.03 MB | 1 年前3
共 25 条
- 1
- 2
- 3













