积分充值
 首页
前端开发
AngularDartElectronFlutterHTML/CSSJavaScriptReactSvelteTypeScriptVue.js构建工具
后端开发
.NetC#C++C语言DenoffmpegGoIdrisJavaJuliaKotlinLeanMakefilenimNode.jsPascalPHPPythonRISC-VRubyRustSwiftUML其它语言区块链开发测试微服务敏捷开发架构设计汇编语言
数据库
Apache DorisApache HBaseCassandraClickHouseFirebirdGreenplumMongoDBMySQLPieCloudDBPostgreSQLRedisSQLSQLiteTiDBVitess数据库中间件数据库工具数据库设计
系统运维
AndroidDevOpshttpdJenkinsLinuxPrometheusTraefikZabbix存储网络与安全
云计算&大数据
Apache APISIXApache FlinkApache KarafApache KyuubiApache OzonedaprDockerHadoopHarborIstioKubernetesOpenShiftPandasrancherRocketMQServerlessService MeshVirtualBoxVMWare云原生CNCF机器学习边缘计算
综合其他
BlenderGIMPKiCadKritaWeblate产品与服务人工智能亿图数据可视化版本控制笔试面试
文库资料
前端
AngularAnt DesignBabelBootstrapChart.jsCSS3EchartsElectronHighchartsHTML/CSSHTML5JavaScriptJerryScriptJestReactSassTypeScriptVue前端工具小程序
后端
.NETApacheC/C++C#CMakeCrystalDartDenoDjangoDubboErlangFastifyFlaskGinGoGoFrameGuzzleIrisJavaJuliaLispLLVMLuaMatplotlibMicronautnimNode.jsPerlPHPPythonQtRPCRubyRustR语言ScalaShellVlangwasmYewZephirZig算法
移动端
AndroidAPP工具FlutterFramework7HarmonyHippyIoniciOSkotlinNativeObject-CPWAReactSwiftuni-appWeex
数据库
ApacheArangoDBCassandraClickHouseCouchDBCrateDBDB2DocumentDBDorisDragonflyDBEdgeDBetcdFirebirdGaussDBGraphGreenPlumHStreamDBHugeGraphimmudbIndexedDBInfluxDBIoTDBKey-ValueKitDBLevelDBM3DBMatrixOneMilvusMongoDBMySQLNavicatNebulaNewSQLNoSQLOceanBaseOpenTSDBOracleOrientDBPostgreSQLPrestoDBQuestDBRedisRocksDBSequoiaDBServerSkytableSQLSQLiteTiDBTiKVTimescaleDBYugabyteDB关系型数据库数据库数据库ORM数据库中间件数据库工具时序数据库
云计算&大数据
ActiveMQAerakiAgentAlluxioAntreaApacheApache APISIXAPISIXBFEBitBookKeeperChaosChoerodonCiliumCloudStackConsulDaprDataEaseDC/OSDockerDrillDruidElasticJobElasticSearchEnvoyErdaFlinkFluentGrafanaHadoopHarborHelmHudiInLongKafkaKnativeKongKubeCubeKubeEdgeKubeflowKubeOperatorKubernetesKubeSphereKubeVelaKumaKylinLibcloudLinkerdLonghornMeiliSearchMeshNacosNATSOKDOpenOpenEBSOpenKruiseOpenPitrixOpenSearchOpenStackOpenTracingOzonePaddlePaddlePolicyPulsarPyTorchRainbondRancherRediSearchScikit-learnServerlessShardingSphereShenYuSparkStormSupersetXuperChainZadig云原生CNCF人工智能区块链数据挖掘机器学习深度学习算法工程边缘计算
UI&美工&设计
BlenderKritaSketchUI设计
网络&系统&运维
AnsibleApacheAWKCeleryCephCI/CDCurveDevOpsGoCDHAProxyIstioJenkinsJumpServerLinuxMacNginxOpenRestyPrometheusServertraefikTrafficUnixWindowsZabbixZipkin安全防护系统内核网络运维监控
综合其它
文章资讯
 上传文档  发布文章  登录账户
IT文库
  • 综合
  • 文档
  • 文章

无数据

分类

全部云计算&大数据(19)机器学习(19)

语言

全部英语(10)中文(简体)(9)

格式

全部PDF文档 PDF(19)
 
本次搜索耗时 0.060 秒,为您找到相关结果约 19 个.
  • 全部
  • 云计算&大数据
  • 机器学习
  • 全部
  • 英语
  • 中文(简体)
  • 全部
  • PDF文档 PDF
  • 默认排序
  • 最新排序
  • 页数排序
  • 大小排序
  • 全部时间
  • 最近一天
  • 最近一周
  • 最近一个月
  • 最近三个月
  • 最近半年
  • 最近一年
  • pdf文档 keras tutorial

    through the installation of Keras, basics of deep learning, Keras models, Keras layers, Keras modules and finally conclude with some real-time applications. Audience This tutorial is prepared ... 18 Core Modules ................................................................................................................................................ 19 6. Keras ― Modules ......... ................................................................................ 20 Available modules ................................................................................................
    0 码力 | 98 页 | 1.57 MB | 1 年前
    3
  • pdf文档 PyTorch Release Notes

    cuBLAS 12.1.3.1 ‣ NVIDIA cuDNN 8.9.3 ‣ NVIDIA NCCL 2.18.3 ‣ NVIDIA RAPIDS™ 23.06 ‣ Apex ‣ rdma-core 39.0 ‣ NVIDIA HPC-X 2.15 ‣ OpenMPI 4.1.4+ ‣ GDRCopy 2.3 ‣ TensorBoard 2.9.0 ‣ Nsight Compute with lower memory utilization. Transformer Engine also includes a collection of highly optimized modules for popular Transformer architectures and an automatic mixed precision-like API that can be used For more information about AMP, see the Training With Mixed Precision Guide. Tensor Core Examples The tensor core examples provided in GitHub and NGC focus on achieving the best performance and convergence
    0 码力 | 365 页 | 2.94 MB | 1 年前
    3
  • pdf文档 AI大模型千问 qwen 中文文档

    class LoraArguments: lora_r: int = 64 lora_alpha: int = 16 lora_dropout: float = 0.05 lora_target_modules: List[str] = field( default_factory=lambda: [ "q_proj", "k_proj", "v_proj", "o_proj", "up_proj" lora_alpha: the alpha value for LoRA; • lora_dropout: the dropout rate for LoRA; • lora_target_modules: the target modules for LoRA. By default we tune all linear layers; • lora_weight_path: the path to the weight lora_config = LoraConfig( r=lora_args.lora_r, lora_alpha=lora_args.lora_alpha, target_modules=lora_args.lora_target_modules, lora_dropout=lora_args.lora_dropout, bias=lora_args.lora_bias, task_type="CAUSAL_LM"
    0 码力 | 56 页 | 835.78 KB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 7 - Automation

    HyperBand, the recommended factor is 3. We will use the same. Now, let's go on and load the required modules and the dataset. import tensorflow as tf import tensorflow_datasets as tfds import keras_tuner as dropout_rate=DROPOUT_RATE): # Initalize the core model core_args = dict(input_shape=(IMG_SIZE, IMG_SIZE, 3), include_top=False) core = apps.resnet50.ResNet50(**core_args) core.trainable = False # Setup the top Lambda(lambda x: tf.cast(x, tf.float32)), layers.Lambda(lambda x: apps.resnet.preprocess_input(x)), core, layers.Flatten(), layers.Dropout(dropout_rate), layers.Dense(NUM_CLASSES, activation='softmax')
    0 码力 | 33 页 | 2.48 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 3 - Learning Techniques

    create_model(): # Initialize the core model core_args = dict(input_shape=(IMG_SIZE, IMG_SIZE, 3), include_top=False) core = apps.resnet50.ResNet50(**core_args) core.trainable = False # Create the full full model with input, preprocessing, core and softmax layers. model = tf.keras.Sequential([ layers.Input([IMG_SIZE, IMG_SIZE, 3], dtype = tf.uint8), layers.Lambda(lambda x: tf.cast(x, tf.float32)) float32)), layers.Lambda(lambda x: apps.resnet.preprocess_input(x)), core, layers.Flatten(), layers.Dropout(DROPOUT_RATE), layers.Dense(NUM_CLASSES, activation='softmax') ]) adam = optimizers.Adam(learn
    0 码力 | 56 页 | 18.93 MB | 1 年前
    3
  • pdf文档 《Efficient Deep Learning Book》[EDL] Chapter 4 - Efficient Architectures

    the hashing trick. It helps to reduce the vocabulary with little or no performance trade-off. The core idea of the hashing trick is as follows: 1. Choose the desired vocabulary size N, and the number entire Jupyter notebook is here for you to experiment with. We begin with loading the necessary modules. The Oxford-IIIT dataset is available through the tensorflow_datasets package. We apply the standard
    0 码力 | 53 页 | 3.92 MB | 1 年前
    3
  • pdf文档 动手学深度学习 v2.0

    display from matplotlib import pyplot as plt from matplotlib_inline import backend_inline d2l = sys.modules[__name__] 本书中的大部分代码都是基于PyTorch的。PyTorch是一个开源的深度学习框架,在研究界非常受欢迎。本书 中的所有代码都在最新版本的PyTorch下通过了测试。但是 MF (Intel 80186) 1990 10 K (光学字符识别) 10 MB 10 MF (Intel 80486) 2000 10 M (网页) 100 MB 1 GF (Intel Core) 2010 10 G (广告) 1 GB 1 TF (Nvidia C2050) 2020 1 T (社交网络) 100 GB 1 PF (Nvidia DGX‐2) 很明显,随机存取存储 module in enumerate(args): # 这里,module是Module子类的一个实例。我们把它保存在'Module'类的成员 # 变量_modules中。_module的类型是OrderedDict self._modules[str(idx)] = module def forward(self, X): # OrderedDict保证了按照成员添加的顺序遍历它们 (continues
    0 码力 | 797 页 | 29.45 MB | 1 年前
    3
  • pdf文档 深度学习与PyTorch入门实战 - 43. nn.Module

    ▪ Conv2d ▪ ConvTransposed2d ▪ Dropout ▪ etc. 2. Container ▪ net(x) 3. parameters 4. modules ▪ modules: all nodes ▪ children: direct children 5. to(device) 6. save and load 7. train/test 8
    0 码力 | 16 页 | 1.14 MB | 1 年前
    3
  • pdf文档 PyTorch Tutorial

    Sample Code in practice Complex Models • Complex Model Class • Predefined 'layer' modules • 'Sequential' layer modules Dataset • Dataset • In PyTorch, a dataset is represented by a regular Python class
    0 码力 | 38 页 | 4.09 MB | 1 年前
    3
  • pdf文档 全连接神经网络实战. pytorch 版

    的训练带来极大好处。 在 NeuralNetwork 内部定义函数: def weight_init ( s e l f ) : #遍 历 网 络 的 每 一 层 fo r m in s e l f . modules () : #如 果 该 层 是 线 性 连 接 层 i f i s i n s t a n c e (m, nn . Linear ) : print (m. weight . shape 8) , nn .ReLU() , nn . Linear (8 , 4) , ) def weight_init ( s e l f ) : fo r m in s e l f . modules () : i f i s i n s t a n c e (m, nn . Linear ) : m. weight . data . normal_ ( 0 . 0 , 1.0)#. f
    0 码力 | 29 页 | 1.40 MB | 1 年前
    3
共 19 条
  • 1
  • 2
前往
页
相关搜索词
kerastutorialPyTorchReleaseNotesAI模型千问qwen中文文档EfficientDeepLearningBookEDLChapterAutomationTechniquesArchitectures动手深度学习v2入门实战43nnModuleTutorial连接神经网络神经网神经网络pytorch
IT文库
关于我们 文库协议 联系我们 意见反馈 免责声明
本站文档数据由用户上传或本站整理自互联网,不以营利为目的,供所有人免费下载和学习使用。如侵犯您的权益,请联系我们进行删除。
IT文库 ©1024 - 2025 | 站点地图
Powered By MOREDOC AI v3.3.0-beta.70
  • 关注我们的公众号【刻舟求荐】,给您不一样的精彩
    关注我们的公众号【刻舟求荐】,给您不一样的精彩