 OpenShift Container Platform 4.4 构建(build)OpenShift Container Platform 4.4 构建(build) 在 OpenShift Container Platform 中执行构建并与之交互 Last Updated: 2021-03-11 OpenShift Container Platform 4.4 构建(build) 在 OpenShift Container Platform 中执行构建并与之交互 . . . . . . . . . . . . . . . . . . . . . . . . . 目 目录 录 第 第 1 章 章 理解 理解镜 镜像 像构 构建 建 1.1. 构建(BUILD) 第 第 2 章 章 了解 了解构 构建配置 建配置 2.1. BUILDCONFIG 第 第 3 章 章 创 创建 建构 构建 建输 输入 入 3.1. 构建输入 3.2. DOCKERFILE 87 87 87 88 88 89 93 OpenShift Container Platform 4.4 构 构建( 建(build) ) 2 目 目录 录 3 第 1 章 理解镜像构建 1.1. 构建(BUILD) 构建 (build)是将输入参数转换为结果对象的过程。此过程最常用于将输入参数或源代码转换为可运行的镜 像。BuildConfig 对象是整个构建过程的定义。0 码力 | 101 页 | 1.12 MB | 1 年前3 OpenShift Container Platform 4.4 构建(build)OpenShift Container Platform 4.4 构建(build) 在 OpenShift Container Platform 中执行构建并与之交互 Last Updated: 2021-03-11 OpenShift Container Platform 4.4 构建(build) 在 OpenShift Container Platform 中执行构建并与之交互 . . . . . . . . . . . . . . . . . . . . . . . . . 目 目录 录 第 第 1 章 章 理解 理解镜 镜像 像构 构建 建 1.1. 构建(BUILD) 第 第 2 章 章 了解 了解构 构建配置 建配置 2.1. BUILDCONFIG 第 第 3 章 章 创 创建 建构 构建 建输 输入 入 3.1. 构建输入 3.2. DOCKERFILE 87 87 87 88 88 89 93 OpenShift Container Platform 4.4 构 构建( 建(build) ) 2 目 目录 录 3 第 1 章 理解镜像构建 1.1. 构建(BUILD) 构建 (build)是将输入参数转换为结果对象的过程。此过程最常用于将输入参数或源代码转换为可运行的镜 像。BuildConfig 对象是整个构建过程的定义。0 码力 | 101 页 | 1.12 MB | 1 年前3
 Using Istio to Build the Next 5G PlatformUsing Istio to Build the Next 5G Platform David Lenrow Open Source Service Mesh Evangelist Neeraj Poddar Co-founder & Chief Architect, Aspen Mesh February 22, 2021 2 ©2021 Aspen Mesh. All rights reserved Observability, Debugging Uniform metrics and tracing for all CNF traffic Enforcement Primitives to Build Zero Trust Strong identity for users, workloads, devices, etc. Encrypting inter-CNF traffic via0 码力 | 18 页 | 3.79 MB | 1 年前3 Using Istio to Build the Next 5G PlatformUsing Istio to Build the Next 5G Platform David Lenrow Open Source Service Mesh Evangelist Neeraj Poddar Co-founder & Chief Architect, Aspen Mesh February 22, 2021 2 ©2021 Aspen Mesh. All rights reserved Observability, Debugging Uniform metrics and tracing for all CNF traffic Enforcement Primitives to Build Zero Trust Strong identity for users, workloads, devices, etc. Encrypting inter-CNF traffic via0 码力 | 18 页 | 3.79 MB | 1 年前3
 OpenShift Container Platform 4.13 CI/CD1. OPENSHIFT 构建 1.2. OPENSHIFT PIPELINES 1.3. OPENSHIFT GITOPS 1.4. JENKINS 第 第 2 章 章 构 构建( 建(BUILD) ) 2.1. 理解镜像构建 2.2. 了解构建配置 2.3. 创建构建输入 2.4. 管理构建输出 2.5. 使用构建策略 2.6. 使用 BUILDAH 自定义镜像构建 2.7. 执行和配置基本构建 OpenShift 上部署。 第 第 1 章 章 OPENSHIFT CONTAINER PLATFORM CI/CD 概述 概述 3 第 2 章 构建(BUILD) 2.1. 理解镜像构建 2.1.1. Builds 构建 (build) 是将输入参数转换为结果对象的过程。此过程最常用于将输入参数或源代码转换为可运行的 镜像。BuildConfig 对象是整个构建过程的定义。 OpenShift 在每次容器镜像标签或源代码改变时产生新的构建: BuildConfig 对 对象定 象定义 义 kind: BuildConfig apiVersion: build.openshift.io/v1 metadata: name: "ruby-sample-build" 1 spec: runPolicy: "Serial" 2 triggers: 3 - type: "GitHub"0 码力 | 129 页 | 1.37 MB | 1 年前3 OpenShift Container Platform 4.13 CI/CD1. OPENSHIFT 构建 1.2. OPENSHIFT PIPELINES 1.3. OPENSHIFT GITOPS 1.4. JENKINS 第 第 2 章 章 构 构建( 建(BUILD) ) 2.1. 理解镜像构建 2.2. 了解构建配置 2.3. 创建构建输入 2.4. 管理构建输出 2.5. 使用构建策略 2.6. 使用 BUILDAH 自定义镜像构建 2.7. 执行和配置基本构建 OpenShift 上部署。 第 第 1 章 章 OPENSHIFT CONTAINER PLATFORM CI/CD 概述 概述 3 第 2 章 构建(BUILD) 2.1. 理解镜像构建 2.1.1. Builds 构建 (build) 是将输入参数转换为结果对象的过程。此过程最常用于将输入参数或源代码转换为可运行的 镜像。BuildConfig 对象是整个构建过程的定义。 OpenShift 在每次容器镜像标签或源代码改变时产生新的构建: BuildConfig 对 对象定 象定义 义 kind: BuildConfig apiVersion: build.openshift.io/v1 metadata: name: "ruby-sample-build" 1 spec: runPolicy: "Serial" 2 triggers: 3 - type: "GitHub"0 码力 | 129 页 | 1.37 MB | 1 年前3
 OpenShift Container Platform 4.10 CLI 工具RESTARTS AGE IP NODE NOMINATED NODE cakephp-ex-1-build 0/1 Completed 0 5m45s 10.131.0.10 ip-10-0-141-74.ec2.internal OpenShift Container Platform 4.10 CLI 工具RESTARTS AGE IP NODE NOMINATED NODE cakephp-ex-1-build 0/1 Completed 0 5m45s 10.131.0.10 ip-10-0-141-74.ec2.internal- cakephp-ex-1-deploy 息: 示例: 示例:查 查看 看 Pod 资 资源的文档 源的文档 输 输出示例 出示例 OpenShift Client This client helps you develop, build, deploy, and run your applications on any OpenShift or Kubernetes compatible platform. It also includes autoscale deployment foo --min=2 --max=10 第 第 2 章 章 OPENSHIFT CLI (OC) 25 2.5.1.12. oc cancel-build 取消正在运行、待处理或新的构建 用法示例 用法示例 2.5.1.13. oc cluster-info 显示集群信息 用法示例 用法示例 2.5.1.14. oc cluster-info 0 码力 | 120 页 | 1.04 MB | 1 年前3
 OpenShift Container Platform 4.13 CLI 工具RESTARTS AGE IP NODE NOMINATED NODE cakephp-ex-1-build 0/1 Completed 0 5m45s 10.131.0.10 ip-10-0-141-74.ec2.internal OpenShift Container Platform 4.13 CLI 工具RESTARTS AGE IP NODE NOMINATED NODE cakephp-ex-1-build 0/1 Completed 0 5m45s 10.131.0.10 ip-10-0-141-74.ec2.internal- cakephp-ex-1-deploy true ConfigMap ... $ oc help OpenShift Client This client helps you develop, build, deploy, and run your applications on any OpenShift or Kubernetes compatible platform. It also includes 则 用法示例 用法示例 2.7.1.11. oc autoscale 自动缩放部署配置、部署、副本集、有状态集或复制控制器 用法示例 用法示例 2.7.1.12. oc cancel-build 取消正在运行、待处理或新的构建 用法示例 用法示例 2.7.1.13. oc cluster-info 显示集群信息 用法示例 用法示例 # Reconcile RBAC resources 0 码力 | 128 页 | 1.11 MB | 1 年前3
 OpenShift Container Platform 4.8 CLI 工具RESTARTS AGE IP NODE NOMINATED NODE cakephp-ex-1-build 0/1 Completed 0 5m45s 10.131.0.10 ip-10-0-141-74.ec2.internal OpenShift Container Platform 4.8 CLI 工具RESTARTS AGE IP NODE NOMINATED NODE cakephp-ex-1-build 0/1 Completed 0 5m45s 10.131.0.10 ip-10-0-141-74.ec2.internal- cakephp-ex-1-deploy 息: 示例: 示例:查 查看 看 Pod 资 资源的文档 源的文档 输 输出示例 出示例 OpenShift Client This client helps you develop, build, deploy, and run your applications on any OpenShift or Kubernetes compatible platform. It also includes autoscale deployment foo --min=2 --max=10 第 第 2 章 章 OPENSHIFT CLI (OC) 25 2.5.1.12. oc cancel-build 取消正在运行、待处理或新的构建 用法示例 用法示例 2.5.1.13. oc cluster-info 显示集群信息 用法示例 用法示例 2.5.1.14. oc cluster-info 0 码力 | 152 页 | 1.24 MB | 1 年前3
 Apache Kyuubi 1.7.0-rc1 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 206 页 | 3.78 MB | 1 年前3 Apache Kyuubi 1.7.0-rc1 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 206 页 | 3.78 MB | 1 年前3
 Apache Kyuubi 1.7.0 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 206 页 | 3.78 MB | 1 年前3 Apache Kyuubi 1.7.0 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 206 页 | 3.78 MB | 1 年前3
 Apache Kyuubi 1.7.0-rc0 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 210 页 | 3.79 MB | 1 年前3 Apache Kyuubi 1.7.0-rc0 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 210 页 | 3.79 MB | 1 年前3
 Apache Kyuubi 1.7.3 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 211 页 | 3.79 MB | 1 年前3 Apache Kyuubi 1.7.3 Documentationa child’s play to combine some of the components above to build a modern data stack. For example, you can use Kyuubi, Spark and Iceberg to build and manage Data Lakehouse with pure SQL for both data processing lake in cloud storage or an on-prem HDFS cluster. • Lakehouse formation and analytics – Easily build an ACID table storage layer via Hudi, Iceberg, or/and Delta Lake. • Logical data warehouse – Provide sheet can be used individually or all together. For example, you can use Kyuubi, Spark and Flink to build a streaming data warehouse. And then, you can use Zookeeper to enable the load balancing for high0 码力 | 211 页 | 3.79 MB | 1 年前3
共 349 条
- 1
- 2
- 3
- 4
- 5
- 6
- 35














 
 