向量数据量milvus k8s helm 对接外部安装部署流程

avatar
作者
筋斗云
阅读量:2

前情概要:历经了太多的坑,从简单的到困难的,该文章主要是为大家尽可能的展现安装部署流程中遇见的坑!
如果2024年7月15日17:13:41 你处在这个时间阶段 附近,你会发现docker下载镜像失败! 这个问题,没有办法,请使用魔法

官方部署网址:https://milvus.io/docs/install_cluster-helm.md
1.如果你想要直接部署,不对接外部组件,直接使用在线部署,当前要注意上面的问题:使用魔法先把需要的镜像下载下来!
镜像如下:

milvusdb/milvus:  milvusdb/milvus-config-tool: docker.io/milvusdb/etcd: zilliz/attu: 

value.yaml

## Enable or disable Milvus Cluster mode cluster:   enabled: true  image:   all:     repository: milvusdb/milvus     tag: v2.4.5     pullPolicy: IfNotPresent     ## Optionally specify an array of imagePullSecrets.     ## Secrets must be manually created in the namespace.     ## ref: https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry/     ##     # pullSecrets:     #   - myRegistryKeySecretName   tools:     repository: milvusdb/milvus-config-tool     tag: v0.1.2     pullPolicy: IfNotPresent  # Global node selector # If set, this will apply to all milvus components # Individual components can be set to a different node selector nodeSelector: {}  # Global tolerations # If set, this will apply to all milvus components # Individual components can be set to a different tolerations tolerations: []  # Global affinity # If set, this will apply to all milvus components # Individual components can be set to a different affinity affinity: {}  # Global labels and annotations # If set, this will apply to all milvus components labels: {} annotations: {}  # Extra configs for milvus.yaml # If set, this config will merge into milvus.yaml # Please follow the config structure in the milvus.yaml # at https://github.com/milvus-io/milvus/blob/master/configs/milvus.yaml # Note: this config will be the top priority which will override the config # in the image and helm chart. extraConfigFiles:   user.yaml: |+     #    For example enable rest http for milvus proxy     #    proxy:     #      http:     #        enabled: true     #      maxUserNum: 100     #      maxRoleNum: 10     ##  Enable tlsMode and set the tls cert and key     #  tls:     #    serverPemPath: /etc/milvus/certs/tls.crt     #    serverKeyPath: /etc/milvus/certs/tls.key     #   common:     #     security:     #       tlsMode: 1  ## Expose the Milvus service to be accessed from outside the cluster (LoadBalancer service). ## or access it from within the cluster (ClusterIP service). Set the service type and the port to serve it. ## ref: http://kubernetes.io/docs/user-guide/services/ ## service:   type: NodePort   port: 19530   portName: milvus   nodePort: ""   annotations: {}   labels: {}    ## List of IP addresses at which the Milvus service is available   ## Ref: https://kubernetes.io/docs/user-guide/services/#external-ips   ##   externalIPs: []   #   - externalIp1    # LoadBalancerSourcesRange is a list of allowed CIDR values, which are combined with ServicePort to   # set allowed inbound rules on the security group assigned to the master load balancer   loadBalancerSourceRanges:   - 0.0.0.0/0   # Optionally assign a known public LB IP   # loadBalancerIP: 1.2.3.4  ingress:   enabled: false   annotations:     # Annotation example: set nginx ingress type     # kubernetes.io/ingress.class: nginx     nginx.ingress.kubernetes.io/backend-protocol: GRPC     nginx.ingress.kubernetes.io/listen-ports-ssl: '[19530]'     nginx.ingress.kubernetes.io/proxy-body-size: 4m     nginx.ingress.kubernetes.io/ssl-redirect: "true"   labels: {}   rules:     - host: "milvus-example.local"       path: "/"       pathType: "Prefix"     # - host: "milvus-example2.local"     #   path: "/otherpath"     #   pathType: "Prefix"   tls: []   #  - secretName: chart-example-tls   #    hosts:   #      - milvus-example.local  serviceAccount:   create: false   name:   annotations:   labels:  metrics:   enabled: true    serviceMonitor:     # Set this to `true` to create ServiceMonitor for Prometheus operator     enabled: false     interval: "30s"     scrapeTimeout: "10s"     # Additional labels that can be used so ServiceMonitor will be discovered by Prometheus     additionalLabels: {}  livenessProbe:   enabled: true   initialDelaySeconds: 90   periodSeconds: 30   timeoutSeconds: 5   successThreshold: 1   failureThreshold: 5  readinessProbe:   enabled: true   initialDelaySeconds: 90   periodSeconds: 10   timeoutSeconds: 5   successThreshold: 1   failureThreshold: 5  log:   level: "info"   file:     maxSize: 300    # MB     maxAge: 10    # day     maxBackups: 20   format: "text"    # text/json    persistence:     mountPath: "/milvus/logs"     ## If true, create/use a Persistent Volume Claim     ## If false, use emptyDir     ##     enabled: false     annotations:       helm.sh/resource-policy: keep     persistentVolumeClaim:       existingClaim: ""       ## Milvus Logs Persistent Volume Storage Class       ## If defined, storageClassName: <storageClass>       ## If set to "-", storageClassName: "", which disables dynamic provisioning       ## If undefined (the default) or set to null, no storageClassName spec is       ##   set, choosing the default provisioner.       ## ReadWriteMany access mode required for milvus cluster.       ##       storageClass:       accessModes: ReadWriteMany       size: 10Gi       subPath: ""  ## Heaptrack traces all memory allocations and annotates these events with stack traces. ## See more: https://github.com/KDE/heaptrack ## Enable heaptrack in production is not recommended. heaptrack:   image:     repository: milvusdb/heaptrack     tag: v0.1.0     pullPolicy: IfNotPresent  standalone:   replicas: 1  # Run standalone mode with replication disabled   resources: {}   # Set local storage size in resources   # resources:   #   limits:   #     ephemeral-storage: 100Gi   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   disk:     enabled: true     size:       enabled: false  # Enable local storage size limit   profiling:     enabled: false  # Enable live profiling    ## Default message queue for milvus standalone   ## Supported value: rocksmq, natsmq, pulsar and kafka   messageQueue: rocksmq   persistence:     mountPath: "/var/lib/milvus"     ## If true, alertmanager will create/use a Persistent Volume Claim     ## If false, use emptyDir     ##     enabled: true     annotations:       helm.sh/resource-policy: keep     persistentVolumeClaim:       existingClaim: ""       ## Milvus Persistent Volume Storage Class       ## If defined, storageClassName: <storageClass>       ## If set to "-", storageClassName: "", which disables dynamic provisioning       ## If undefined (the default) or set to null, no storageClassName spec is       ##   set, choosing the default provisioner.       ##       storageClass: "csi-driver-s3"       accessModes: ReadWriteOnce       size: 50Gi       subPath: ""  proxy:   enabled: true   # You can set the number of replicas to -1 to remove the replicas field in case you want to use HPA   replicas: 1   resources: {}   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   http:     enabled: true  # whether to enable http rest server     debugMode:       enabled: false   # Mount a TLS secret into proxy pod   tls:     enabled: false ## when enabling proxy.tls, all items below should be uncommented and the key and crt values should be populated. #    enabled: true #    secretName: milvus-tls ## expecting base64 encoded values here: i.e. $(cat tls.crt | base64 -w 0) and $(cat tls.key | base64 -w 0) #    key: LS0tLS1CRUdJTiBQU--REDUCT #    crt: LS0tLS1CRUdJTiBDR--REDUCT #  volumes: #  - secret: #      secretName: milvus-tls #    name: milvus-tls #  volumeMounts: #  - mountPath: /etc/milvus/certs/ #    name: milvus-tls   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}  rootCoordinator:   enabled: true   # You can set the number of replicas greater than 1, only if enable active standby   replicas: 1  # Run Root Coordinator mode with replication disabled   resources: {}   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   activeStandby:     enabled: true  # Enable active-standby when you set multiple replicas for root coordinator   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}    service:     port: 53100     annotations: {}     labels: {}     clusterIP: ""  queryCoordinator:   enabled: true   # You can set the number of replicas greater than 1, only if enable active standby   replicas: 1  # Run Query Coordinator mode with replication disabled   resources: {}   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   activeStandby:     enabled: true  # Enable active-standby when you set multiple replicas for query coordinator   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}    service:     port: 19531     annotations: {}     labels: {}     clusterIP: ""  queryNode:   enabled: true   # You can set the number of replicas to -1 to remove the replicas field in case you want to use HPA   replicas: 1   resources: {}   # Set local storage size in resources   # resources:   #   limits:   #     ephemeral-storage: 100Gi   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   disk:     enabled: true  # Enable querynode load disk index, and search on disk index     size:       enabled: false  # Enable local storage size limit   profiling:     enabled: false  # Enable live profiling   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}  indexCoordinator:   enabled: true   # You can set the number of replicas greater than 1, only if enable active standby   replicas: 1   # Run Index Coordinator mode with replication disabled   resources: {}   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   activeStandby:     enabled: true  # Enable active-standby when you set multiple replicas for index coordinator   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}    service:     port: 31000     annotations: {}     labels: {}     clusterIP: ""  indexNode:   enabled: true   # You can set the number of replicas to -1 to remove the replicas field in case you want to use HPA   replicas: 1   resources: {}   # Set local storage size in resources   # limits:   #    ephemeral-storage: 100Gi   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   disk:     enabled: true  # Enable index node build disk vector index     size:       enabled: false  # Enable local storage size limit   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}  dataCoordinator:   enabled: true   # You can set the number of replicas greater than 1, only if enable active standby   replicas: 1           # Run Data Coordinator mode with replication disabled   resources: {}   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   activeStandby:     enabled: true  # Enable active-standby when you set multiple replicas for data coordinator   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}    service:     port: 13333     annotations: {}     labels: {}     clusterIP: ""  dataNode:   enabled: true   # You can set the number of replicas to -1 to remove the replicas field in case you want to use HPA   replicas: 1   resources: {}   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}  ## mixCoordinator contains all coord ## If you want to use mixcoord, enable this and disable all of other coords mixCoordinator:   enabled: false   # You can set the number of replicas greater than 1, only if enable active standby   replicas: 1           # Run Mixture Coordinator mode with replication disabled   resources: {}   nodeSelector: {}   affinity: {}   tolerations: []   extraEnv: []   heaptrack:     enabled: false   profiling:     enabled: false  # Enable live profiling   activeStandby:     enabled: true  # Enable active-standby when you set multiple replicas for Mixture coordinator   # Deployment strategy, default is RollingUpdate   # Ref: https://kubernetes.io/docs/concepts/workloads/controllers/deployment/#rolling-update-deployment   strategy: {}    service:     annotations: {}     labels: {}     clusterIP: ""  attu:   enabled: true   name: attu   image:     repository: zilliz/attu     tag: v2.3.10     pullPolicy: IfNotPresent   service:     annotations: {}     labels: {}     type: NodePort     port: 3000     # loadBalancerIP: ""   resources: {}   podLabels: {}   ingress:     enabled: false     annotations: {}     # Annotation example: set nginx ingress type     # kubernetes.io/ingress.class: nginx     labels: {}     hosts:       - milvus-attu.local     tls: []     #  - secretName: chart-attu-tls     #    hosts:     #      - milvus-attu.local   ## Configuration values for the minio dependency ## ref: https://github.com/zilliztech/milvus-helm/blob/master/charts/minio/README.md ##  minio:   enabled: false   name: minio   mode: distributed   image:     tag: "RELEASE.2023-03-20T20-16-18Z"     pullPolicy: IfNotPresent   accessKey: minioadmin   secretKey: minioadmin   existingSecret: ""   bucketName: "milvus-bucket"   rootPath: file   useIAM: false   iamEndpoint: ""   region: ""   useVirtualHost: false   podDisruptionBudget:     enabled: false   resources:     requests:       memory: 2Gi    service:     type: ClusterIP     port: 9000    persistence:     enabled: true     existingClaim: ""     storageClass: "csi-driver-s3"     accessMode: ReadWriteOnce     size: 500Gi    livenessProbe:     enabled: true     initialDelaySeconds: 5     periodSeconds: 5     timeoutSeconds: 5     successThreshold: 1     failureThreshold: 5    readinessProbe:     enabled: true     initialDelaySeconds: 5     periodSeconds: 5     timeoutSeconds: 1     successThreshold: 1     failureThreshold: 5    startupProbe:     enabled: true     initialDelaySeconds: 0     periodSeconds: 10     timeoutSeconds: 5     successThreshold: 1     failureThreshold: 60  ## Configuration values for the etcd dependency ## ref: https://artifacthub.io/packages/helm/bitnami/etcd ##  etcd:   enabled: false   name: etcd   replicaCount: 3   pdb:     create: false   image:     repository: "milvusdb/etcd"     tag: "3.5.5-r4"     pullPolicy: IfNotPresent    service:     type: ClusterIP     port: 2379     peerPort: 2380    auth:     rbac:       enabled: false    persistence:     enabled: true     storageClass: "csi-driver-s3"     accessMode: ReadWriteOnce     size: 10Gi    ## Change default timeout periods to mitigate zoobie probe process   livenessProbe:     enabled: true     timeoutSeconds: 10    readinessProbe:     enabled: true     periodSeconds: 20     timeoutSeconds: 10    ## Enable auto compaction   ## compaction by every 1000 revision   ##   autoCompactionMode: revision   autoCompactionRetention: "1000"    ## Increase default quota to 4G   ##   extraEnvVars:   - name: ETCD_QUOTA_BACKEND_BYTES     value: "4294967296"   - name: ETCD_HEARTBEAT_INTERVAL     value: "500"   - name: ETCD_ELECTION_TIMEOUT     value: "2500"  ## Configuration values for the pulsar dependency ## ref: https://github.com/apache/pulsar-helm-chart ##  pulsar:   enabled: false   name: pulsar    fullnameOverride: ""   persistence: true    maxMessageSize: "5242880"  # 5 * 1024 * 1024 Bytes, Maximum size of each message in pulsar.    rbac:     enabled: false     psp: false     limit_to_namespace: true    affinity:     anti_affinity: false  ## enableAntiAffinity: no    components:     zookeeper: true     bookkeeper: true     # bookkeeper - autorecovery     autorecovery: true     broker: true     functions: false     proxy: true     toolset: false     pulsar_manager: false    monitoring:     prometheus: false     grafana: false     node_exporter: false     alert_manager: false    images:     broker:       repository: apachepulsar/pulsar       pullPolicy: IfNotPresent       tag: 2.8.2     autorecovery:       repository: apachepulsar/pulsar       tag: 2.8.2       pullPolicy: IfNotPresent     zookeeper:       repository: apachepulsar/pulsar       pullPolicy: IfNotPresent       tag: 2.8.2     bookie:       repository: apachepulsar/pulsar       pullPolicy: IfNotPresent       tag: 2.8.2     proxy:       repository: apachepulsar/pulsar       pullPolicy: IfNotPresent       tag: 2.8.2     pulsar_manager:       repository: apachepulsar/pulsar-manager       pullPolicy: IfNotPresent       tag: v0.1.0    zookeeper:     resources:       requests:         memory: 1024Mi         cpu: 0.3     configData:       PULSAR_MEM: >         -Xms1024m         -Xmx1024m       PULSAR_GC: >          -Dcom.sun.management.jmxremote          -Djute.maxbuffer=10485760          -XX:+ParallelRefProcEnabled          -XX:+UnlockExperimentalVMOptions          -XX:+DoEscapeAnalysis          -XX:+DisableExplicitGC          -XX:+PerfDisableSharedMem          -Dzookeeper.forceSync=no     pdb:       usePolicy: false    bookkeeper:     replicaCount: 3     volumes:       journal:         name: journal         size: 100Gi       ledgers:         name: ledgers         size: 200Gi     resources:       requests:         memory: 2048Mi         cpu: 1     configData:       PULSAR_MEM: >         -Xms4096m         -Xmx4096m         -XX:MaxDirectMemorySize=8192m       PULSAR_GC: >         -Dio.netty.leakDetectionLevel=disabled         -Dio.netty.recycler.linkCapacity=1024         -XX:+UseG1GC -XX:MaxGCPauseMillis=10         -XX:+ParallelRefProcEnabled         -XX:+UnlockExperimentalVMOptions         -XX:+DoEscapeAnalysis         -XX:ParallelGCThreads=32         -XX:ConcGCThreads=32         -XX:G1NewSizePercent=50         -XX:+DisableExplicitGC         -XX:-ResizePLAB         -XX:+ExitOnOutOfMemoryError         -XX:+PerfDisableSharedMem         -XX:+PrintGCDetails       nettyMaxFrameSizeBytes: "104867840"     pdb:       usePolicy: false    broker:     component: broker     podMonitor:       enabled: false     replicaCount: 1     resources:       requests:         memory: 4096Mi         cpu: 1.5     configData:       PULSAR_MEM: >         -Xms4096m         -Xmx4096m         -XX:MaxDirectMemorySize=8192m       PULSAR_GC: >         -Dio.netty.leakDetectionLevel=disabled         -Dio.netty.recycler.linkCapacity=1024         -XX:+ParallelRefProcEnabled         -XX:+UnlockExperimentalVMOptions         -XX:+DoEscapeAnalysis         -XX:ParallelGCThreads=32         -XX:ConcGCThreads=32         -XX:G1NewSizePercent=50         -XX:+DisableExplicitGC         -XX:-ResizePLAB         -XX:+ExitOnOutOfMemoryError       maxMessageSize: "104857600"       defaultRetentionTimeInMinutes: "10080"       defaultRetentionSizeInMB: "-1"       backlogQuotaDefaultLimitGB: "8"       ttlDurationDefaultInSeconds: "259200"       subscriptionExpirationTimeMinutes: "3"       backlogQuotaDefaultRetentionPolicy: producer_exception     pdb:       usePolicy: false    autorecovery:     resources:       requests:         memory: 512Mi         cpu: 1    proxy:     replicaCount: 1     podMonitor:       enabled: false     resources:       requests:         memory: 2048Mi         cpu: 1     service:       type: ClusterIP     ports:       pulsar: 6650     configData:       PULSAR_MEM: >         -Xms2048m -Xmx2048m       PULSAR_GC: >         -XX:MaxDirectMemorySize=2048m       httpNumThreads: "100"     pdb:       usePolicy: false    pulsar_manager:     service:       type: ClusterIP    pulsar_metadata:     component: pulsar-init     image:       # the image used for running `pulsar-cluster-initialize` job       repository: apachepulsar/pulsar       tag: 2.8.2   ## Configuration values for the kafka dependency ## ref: https://artifacthub.io/packages/helm/bitnami/kafka ##  kafka:   enabled: false   name: kafka   replicaCount: 3   image:     repository: bitnami/kafka     tag: 3.1.0-debian-10-r52   ## Increase graceful termination for kafka graceful shutdown   terminationGracePeriodSeconds: "90"   pdb:     create: false    ## Enable startup probe to prevent pod restart during recovering   startupProbe:     enabled: true    ## Kafka Java Heap size   heapOpts: "-Xmx4096m -Xms4096m"   maxMessageBytes: _10485760   defaultReplicationFactor: 3   offsetsTopicReplicationFactor: 3   ## Only enable time based log retention   logRetentionHours: 168   logRetentionBytes: _-1   extraEnvVars:   - name: KAFKA_CFG_MAX_PARTITION_FETCH_BYTES     value: "5242880"   - name: KAFKA_CFG_MAX_REQUEST_SIZE     value: "5242880"   - name: KAFKA_CFG_REPLICA_FETCH_MAX_BYTES     value: "10485760"   - name: KAFKA_CFG_FETCH_MESSAGE_MAX_BYTES     value: "5242880"   - name: KAFKA_CFG_LOG_ROLL_HOURS     value: "24"    persistence:     enabled: true     storageClass:     accessMode: ReadWriteOnce     size: 300Gi    metrics:     ## Prometheus Kafka exporter: exposes complimentary metrics to JMX exporter     kafka:       enabled: false       image:         repository: bitnami/kafka-exporter         tag: 1.4.2-debian-10-r182      ## Prometheus JMX exporter: exposes the majority of Kafkas metrics     jmx:       enabled: false       image:         repository: bitnami/jmx-exporter         tag: 0.16.1-debian-10-r245      ## To enable serviceMonitor, you must enable either kafka exporter or jmx exporter.     ## And you can enable them both     serviceMonitor:       enabled: false    service:     type: ClusterIP     ports:       client: 9092    zookeeper:     enabled: true     replicaCount: 3  ################################### # External S3 # - these configs are only used when `externalS3.enabled` is true ################################### externalS3:   enabled: true   host: "172.20.1.124"   port: "9000"   accessKey: "minioadmin"   secretKey: "minioadmin"   useSSL: false   bucketName: "milvus-dev"   rootPath: ""   useIAM: false   cloudProvider: "aws"   iamEndpoint: ""   region: ""   useVirtualHost: false  ################################### # GCS Gateway # - these configs are only used when `minio.gcsgateway.enabled` is true ################################### externalGcs:   bucketName: ""  ################################### # External etcd # - these configs are only used when `externalEtcd.enabled` is true ################################### externalEtcd:   enabled: true    ## the endpoints of the external etcd   ##   endpoints:     - xxxx:23790  ################################### # External pulsar # - these configs are only used when `externalPulsar.enabled` is true ################################### externalPulsar:   enabled: true   host: "xxx"   port: 30012   maxMessageSize: "5242880"  # 5 * 1024 * 1024 Bytes, Maximum size of each message in pulsar.   tenant: "xx"   namespace: "xxx"   authPlugin: "org.apache.pulsar.client.impl.auth.AuthenticationToken"   authParams: token:"xxx"  ################################### # External kafka # - these configs are only used when `externalKafka.enabled` is true # - note that the following are just examples, you should confirm the #   value of brokerList and mechanisms according to the actual external #   Kafka configuration. E.g. If you select the AWS MSK, the configuration #   should look something like this: #   externalKafka: #     enabled: true #     brokerList: "xxxx:9096" #     securityProtocol: SASL_SSL #     sasl: #       mechanisms: SCRAM-SHA-512 #       password: "xxx" #       username: "xxx" ################################### externalKafka:   enabled: false   brokerList: localhost:9092   securityProtocol: SASL_SSL   sasl:     mechanisms: PLAIN     username: ""     password: ""  

k8s可执行文件milvus_manifest.yaml

--- # Source: milvus/templates/configmap.yaml apiVersion: v1 kind: ConfigMap metadata:   name: my-release-milvus data:   default.yaml: |+     # Copyright (C) 2019-2021 Zilliz. All rights reserved.     #     # Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance     # with the License. You may obtain a copy of the License at     #     # http://www.apache.org/licenses/LICENSE-2.0     #     # Unless required by applicable law or agreed to in writing, software distributed under the License     # is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express     # or implied. See the License for the specific language governing permissions and limitations under the License.          etcd:       endpoints:         - xxxx:23790          metastore:       type: etcd          minio:       address: xxxx       port: 9000       accessKeyID: minioadmin       secretAccessKey: minioadmin       useSSL: false       bucketName: milvus-dev       rootPath:       useIAM: false       cloudProvider: aws       iamEndpoint:       region:       useVirtualHost: false      mq:       type: pulsar      messageQueue: pulsar      pulsar:       address: xxx       port: 6650       maxMessageSize: 5242880       tenant: "my-tenant"       namespace: my-namespace           rootCoord:       address: my-release-milvus-rootcoord       port: 53100       enableActiveStandby: true  # Enable rootcoord active-standby          proxy:       port: 19530       internalPort: 19529          queryCoord:       address: my-release-milvus-querycoord       port: 19531            enableActiveStandby: true  # Enable querycoord active-standby          queryNode:       port: 21123       enableDisk: true # Enable querynode load disk index, and search on disk index          indexCoord:       address: my-release-milvus-indexcoord       port: 31000       enableActiveStandby: true  # Enable indexcoord active-standby          indexNode:       port: 21121       enableDisk: true # Enable index node build disk vector index          dataCoord:       address: my-release-milvus-datacoord       port: 13333       enableActiveStandby: true  # Enable datacoord active-standby          dataNode:       port: 21124          log:       level: info       file:         rootPath: ""         maxSize: 300         maxAge: 10         maxBackups: 20       format: text   user.yaml: |-     #    For example enable rest http for milvus proxy     #    proxy:     #      http:     #        enabled: true     #      maxUserNum: 100     #      maxRoleNum: 10     ##  Enable tlsMode and set the tls cert and key     #  tls:     #    serverPemPath: /etc/milvus/certs/tls.crt     #    serverKeyPath: /etc/milvus/certs/tls.key     #   common:     #     security:     #       tlsMode: 1 --- # Source: milvus/templates/attu-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-attu   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "attu" spec:   type: NodePort   ports:     - name: attu       protocol: TCP       port: 3000       targetPort: 3000   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "attu" --- # Source: milvus/templates/datacoord-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-datacoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "datacoord" spec:   type: ClusterIP   ports:     - name: datacoord       port: 13333       protocol: TCP       targetPort: datacoord     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "datacoord" --- # Source: milvus/templates/datanode-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-datanode   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "datanode" spec:   type: ClusterIP   clusterIP: None   ports:     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "datanode" --- # Source: milvus/templates/indexcoord-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-indexcoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "indexcoord" spec:   type: ClusterIP   ports:     - name: indexcoord       port: 31000       protocol: TCP       targetPort: indexcoord     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "indexcoord" --- # Source: milvus/templates/indexnode-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-indexnode   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "indexnode" spec:   type: ClusterIP   clusterIP: None   ports:     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "indexnode" --- # Source: milvus/templates/querycoord-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-querycoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "querycoord" spec:   type: ClusterIP   ports:     - name: querycoord       port: 19531       protocol: TCP       targetPort: querycoord     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "querycoord" --- # Source: milvus/templates/querynode-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-querynode   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "querynode" spec:   type: ClusterIP   clusterIP: None   ports:     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "querynode" --- # Source: milvus/templates/rootcoord-svc.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus-rootcoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "rootcoord" spec:   type: ClusterIP   ports:     - name: rootcoord       port: 53100       protocol: TCP       targetPort: rootcoord     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "rootcoord" --- # Source: milvus/templates/service.yaml apiVersion: v1 kind: Service metadata:   name: my-release-milvus   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "proxy" spec:   type: NodePort   ports:     - name: milvus       port: 19530       protocol: TCP       targetPort: milvus     - name: metrics       protocol: TCP       port: 9091       targetPort: metrics   selector:     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     component: "proxy" --- # Source: milvus/templates/attu-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-attu   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "attu"      spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "attu"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "attu"              spec:       containers:       - name: attu         image: zilliz/attu:v2.3.10         imagePullPolicy: IfNotPresent         ports:         - name: attu           containerPort: 3000           protocol: TCP         env:         - name: MILVUS_URL           value: http://my-release-milvus:19530         resources:           {} --- # Source: milvus/templates/datacoord-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-datacoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "datacoord"        annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "datacoord"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "datacoord"                annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: datacoord         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "datacoord" ]         env:         ports:           - name: datacoord             containerPort: 13333             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools        volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {} --- # Source: milvus/templates/datanode-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-datanode   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "datanode"         annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "datanode"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "datanode"                annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: datanode         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "datanode" ]         env:         ports:           - name: datanode             containerPort: 21124             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools       volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {} --- # Source: milvus/templates/indexcoord-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-indexcoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "indexcoord"        annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "indexcoord"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "indexcoord"                annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: indexcoord         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "indexcoord" ]         env:         ports:           - name: indexcoord             containerPort: 31000             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools        volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {} --- # Source: milvus/templates/indexnode-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-indexnode   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "indexnode"        annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "indexnode"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "indexnode"                annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: indexnode         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "indexnode" ]         env:         ports:           - name: indexnode             containerPort: 21121             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools         - mountPath: /var/lib/milvus/data           name: disk        volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {}       - name: disk         emptyDir: {} --- # Source: milvus/templates/proxy-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-proxy   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "proxy"         annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "proxy"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "proxy"                annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: proxy         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "proxy" ]         env:         ports:           - name: milvus             containerPort: 19530             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools        volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {} --- # Source: milvus/templates/querycoord-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-querycoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "querycoord"        annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "querycoord"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release                  component: "querycoord"       annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: querycoord         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "querycoord" ]         env:         ports:           - name: querycoord             containerPort: 19531             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools        volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {} --- # Source: milvus/templates/querynode-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-querynode   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "querynode"        annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "querynode"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "querynode"                annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: querynode         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "querynode" ]         env:         ports:           - name: querynode             containerPort: 21123             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools         - mountPath: /var/lib/milvus/data           name: disk        volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {}       - name: disk         emptyDir: {} --- # Source: milvus/templates/rootcoord-deployment.yaml apiVersion: apps/v1 kind: Deployment metadata:   name: my-release-milvus-rootcoord   labels:     helm.sh/chart: milvus-4.1.34     app.kubernetes.io/name: milvus     app.kubernetes.io/instance: my-release     app.kubernetes.io/version: "2.4.5"     app.kubernetes.io/managed-by: Helm     component: "rootcoord"        annotations:       spec:   replicas: 1   selector:     matchLabels:       app.kubernetes.io/name: milvus       app.kubernetes.io/instance: my-release       component: "rootcoord"   template:     metadata:       labels:         app.kubernetes.io/name: milvus         app.kubernetes.io/instance: my-release         component: "rootcoord"                annotations:         checksum/config: 4d919a6f7279f31d3f04198e9626ab7a0dec59a9e2d63b9b0758840233e77b8f              spec:       serviceAccountName: default       initContainers:       - name: config         command:         - /cp         - /run-helm.sh,/merge         - /milvus/tools/run-helm.sh,/milvus/tools/merge         image: "milvusdb/milvus-config-tool:v0.1.2"         imagePullPolicy: IfNotPresent         volumeMounts:         - mountPath: /milvus/tools           name: tools       containers:       - name: rootcoord         image: "milvusdb/milvus:v2.4.5"         imagePullPolicy: IfNotPresent         args: [ "/milvus/tools/run-helm.sh", "milvus", "run", "rootcoord" ]         env:         ports:           - name: rootcoord             containerPort: 53100             protocol: TCP           - name: metrics             containerPort: 9091             protocol: TCP         livenessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 30           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         readinessProbe:           httpGet:             path: /healthz             port: metrics           initialDelaySeconds: 90           periodSeconds: 10           timeoutSeconds: 5           successThreshold: 1           failureThreshold: 5         resources:           {}         volumeMounts:         - name: milvus-config           mountPath: /milvus/configs/default.yaml           subPath: default.yaml           readOnly: true         - name: milvus-config           mountPath: /milvus/configs/user.yaml           subPath: user.yaml           readOnly: true         - mountPath: /milvus/tools           name: tools        volumes:       - name: milvus-config         configMap:           name: my-release-milvus       - name: tools         emptyDir: {}  

广告一刻

为您即时展示最新活动产品广告消息,让您随时掌握产品活动新动态!