 OpenAI 《A practical guide to building agents》resisted automation, especially where traditional methods encounter friction: 01 Complex decision-making: Workflows involving nuanced judgment, exceptions, or context-sensitive decisions, for example Difficult-to-maintain rules: Systems that have become unwieldy due to extensive and intricate rulesets, making updates costly or error-prone, for example performing vendor security reviews. 03 Heavy reliance agent consists of three core components: 01 Model The LLM powering the agent’s reasoning and decision-making 02 Tools External functions or APIs the agent can use to take action 03 Instructions Explicit0 码力 | 34 页 | 7.00 MB | 6 月前3 OpenAI 《A practical guide to building agents》resisted automation, especially where traditional methods encounter friction: 01 Complex decision-making: Workflows involving nuanced judgment, exceptions, or context-sensitive decisions, for example Difficult-to-maintain rules: Systems that have become unwieldy due to extensive and intricate rulesets, making updates costly or error-prone, for example performing vendor security reviews. 03 Heavy reliance agent consists of three core components: 01 Model The LLM powering the agent’s reasoning and decision-making 02 Tools External functions or APIs the agent can use to take action 03 Instructions Explicit0 码力 | 34 页 | 7.00 MB | 6 月前3
 Trends Artificial Intelligence
accessible to 5.5B citizens via connected devices; ever-growing digital datasets that have been in the making for over three decades; breakthrough large language models (LLMs) that – in effect – found freedom models, cost-per-token declines, open-source proliferation and chip performance improvements are making new tech advances increasingly more powerful, accessible, and economically viable. OpenAI’s ChatGPT such as math, law, medicine, and history. It measures both factual recall and reasoning ability, making it a standard for assessing general knowledge and problem-solving in large language models. 89.8%0 码力 | 340 页 | 12.14 MB | 4 月前3 Trends Artificial Intelligence
accessible to 5.5B citizens via connected devices; ever-growing digital datasets that have been in the making for over three decades; breakthrough large language models (LLMs) that – in effect – found freedom models, cost-per-token declines, open-source proliferation and chip performance improvements are making new tech advances increasingly more powerful, accessible, and economically viable. OpenAI’s ChatGPT such as math, law, medicine, and history. It measures both factual recall and reasoning ability, making it a standard for assessing general knowledge and problem-solving in large language models. 89.8%0 码力 | 340 页 | 12.14 MB | 4 月前3
 Google 《Prompt Engineering v7》temperature with high certainty. A higher Gemini temperature setting is like a high softmax temperature, making a wider range of temperatures around the selected setting more acceptable. This increased uncertainty sampling implementations will then only consider the most probable token to meet the top-P criteria, making temperature and top-K irrelevant. If you set top-P to 1, any token with a nonzero probability of between system, contextual, and role prompts provides a framework for designing prompts with clear intent, allowing for flexible combinations and making it easier to analyze how each prompt type influences0 码力 | 68 页 | 6.50 MB | 6 月前3 Google 《Prompt Engineering v7》temperature with high certainty. A higher Gemini temperature setting is like a high softmax temperature, making a wider range of temperatures around the selected setting more acceptable. This increased uncertainty sampling implementations will then only consider the most probable token to meet the top-P criteria, making temperature and top-K irrelevant. If you set top-P to 1, any token with a nonzero probability of between system, contextual, and role prompts provides a framework for designing prompts with clear intent, allowing for flexible combinations and making it easier to analyze how each prompt type influences0 码力 | 68 页 | 6.50 MB | 6 月前3
 DeepSeek-V2: A Strong, Economical, and Efficient
Mixture-of-Experts Language Modeltokens. We optimize the attention modules and Feed-Forward Networks (FFNs) within the Trans- former framework (Vaswani et al., 2017) with our proposed Multi-head Latent Attention (MLA) and DeepSeekMoE. (1) Infrastructures DeepSeek-V2 is trained based on the HAI-LLM framework (High-flyer, 2023), an efficient and light-weight training framework developed internally by our engineers. It employs a 16-way zero-bubble in English and Chinese. Our evaluation is based on our internal evaluation framework integrated 13 in our HAI-LLM framework. Included benchmarks are categorized and listed as follows, where underlined0 码力 | 52 页 | 1.23 MB | 1 年前3 DeepSeek-V2: A Strong, Economical, and Efficient
Mixture-of-Experts Language Modeltokens. We optimize the attention modules and Feed-Forward Networks (FFNs) within the Trans- former framework (Vaswani et al., 2017) with our proposed Multi-head Latent Attention (MLA) and DeepSeekMoE. (1) Infrastructures DeepSeek-V2 is trained based on the HAI-LLM framework (High-flyer, 2023), an efficient and light-weight training framework developed internally by our engineers. It employs a 16-way zero-bubble in English and Chinese. Our evaluation is based on our internal evaluation framework integrated 13 in our HAI-LLM framework. Included benchmarks are categorized and listed as follows, where underlined0 码力 | 52 页 | 1.23 MB | 1 年前3
 TVM Meetup: Quantizationdialect© 2019, Amazon Web Services, Inc. or its Affiliates. All rights reserved. TVM Overview Framework Graph Mxnet TF …. parsers Relay Graph Target-independent Relay passes Target-optimized graph .. More targets AutoTVM – Tuning the kernels Optimized Binary Codegen – LLVM, Cuda, C, … Framework Parsers Graph level optimizations Tensor-level optimizations Machine code generation© 2019, Amazon reserved. Quantization Appraoches in TVM Framework FP32 Graph MXNet Parser TF parser …. Relay FP32 Graph Relay Automatic Quantization Relay Int8 Graph Framework Pre-quantized Graph MXNet Parser TF Parser0 码力 | 19 页 | 489.50 KB | 5 月前3 TVM Meetup: Quantizationdialect© 2019, Amazon Web Services, Inc. or its Affiliates. All rights reserved. TVM Overview Framework Graph Mxnet TF …. parsers Relay Graph Target-independent Relay passes Target-optimized graph .. More targets AutoTVM – Tuning the kernels Optimized Binary Codegen – LLVM, Cuda, C, … Framework Parsers Graph level optimizations Tensor-level optimizations Machine code generation© 2019, Amazon reserved. Quantization Appraoches in TVM Framework FP32 Graph MXNet Parser TF parser …. Relay FP32 Graph Relay Automatic Quantization Relay Int8 Graph Framework Pre-quantized Graph MXNet Parser TF Parser0 码力 | 19 页 | 489.50 KB | 5 月前3
 OpenAI - AI in the Enterpriseprocesses, with expert feedback at every step. How it started Morgan Stanley’s first eval focused on making their financial advisors more efficient and effective. The premise was simple: If advisors could world’s No. 1 job site, uses GPT-4o mini to match job seekers to jobs in new ways. The power of why Making great job recommendations to job seekers is only the start of the Indeed experience. They also0 码力 | 25 页 | 9.48 MB | 5 月前3 OpenAI - AI in the Enterpriseprocesses, with expert feedback at every step. How it started Morgan Stanley’s first eval focused on making their financial advisors more efficient and effective. The premise was simple: If advisors could world’s No. 1 job site, uses GPT-4o mini to match job seekers to jobs in new ways. The power of why Making great job recommendations to job seekers is only the start of the Indeed experience. They also0 码力 | 25 页 | 9.48 MB | 5 月前3
 清华大学第二弹:DeepSeek赋能职场作为智能体 ü 角色 ü 功能 ü 技能 ü 约束 ü 工作流程 ü 输出格式 "全维度智能体提示框架" (Comprehensive Agent Prompting Framework, CAP Framework) 核心层: 1.身份定义 (Identity) •角色属性 •专业背景 •交互特征 执行层: 2. 能力矩阵 (Capability Matrix) •功能范围0 码力 | 35 页 | 9.78 MB | 8 月前3 清华大学第二弹:DeepSeek赋能职场作为智能体 ü 角色 ü 功能 ü 技能 ü 约束 ü 工作流程 ü 输出格式 "全维度智能体提示框架" (Comprehensive Agent Prompting Framework, CAP Framework) 核心层: 1.身份定义 (Identity) •角色属性 •专业背景 •交互特征 执行层: 2. 能力矩阵 (Capability Matrix) •功能范围0 码力 | 35 页 | 9.78 MB | 8 月前3
 OctoML OSS 2019 11 8DTYM Runtime send program 较 ,we 人 Interace Optimize TVM operators on microcontrollers by making use of AutoTVM improve0 码力 | 16 页 | 1.77 MB | 5 月前3 OctoML OSS 2019 11 8DTYM Runtime send program 较 ,we 人 Interace Optimize TVM operators on microcontrollers by making use of AutoTVM improve0 码力 | 16 页 | 1.77 MB | 5 月前3
 TVM: Where Are We GoingHardware CuDNN NNPack MKL-DNN Hand optimized Open source, automated end-to- end optimization framework for deep learning.TVM Stack High-Level Differentiable IR Tensor Expression and Optimization0 码力 | 31 页 | 22.64 MB | 5 月前3 TVM: Where Are We GoingHardware CuDNN NNPack MKL-DNN Hand optimized Open source, automated end-to- end optimization framework for deep learning.TVM Stack High-Level Differentiable IR Tensor Expression and Optimization0 码力 | 31 页 | 22.64 MB | 5 月前3
 XDNN TVM - Nov 2019Runtime Image Model Weights Calibration Set Quantizer Compiler Tensor Graph Optimization Framework Tensor Graph to Xilinx Tensor Graph Frontend Deep Learning Frameworks https://github.com/xilinx©0 码力 | 16 页 | 3.35 MB | 5 月前3 XDNN TVM - Nov 2019Runtime Image Model Weights Calibration Set Quantizer Compiler Tensor Graph Optimization Framework Tensor Graph to Xilinx Tensor Graph Frontend Deep Learning Frameworks https://github.com/xilinx©0 码力 | 16 页 | 3.35 MB | 5 月前3
共 11 条
- 1
- 2













