apiVersion: kubekey.kubesphere.io/v1alpha2 kind: Cluster metadata: name: bfs-k8scluster-1 spec: hosts: - {name: bfs-k8smaster-10-2-1-2.hetzner.base.beaconfireinc.com, address: 10.2.1.2, internalAddress: 10.2.1.2, privateKeyPath: "~/.ssh/bfs-ed25519"} - {name: bfs-k8smaster-10-2-1-3.hetzner.base.beaconfireinc.com, address: 10.2.1.3, internalAddress: 10.2.1.3, privateKeyPath: "~/.ssh/bfs-ed25519"} - {name: bfs-k8smaster-10-2-1-4.hetzner.base.beaconfireinc.com, address: 10.2.1.4, internalAddress: 10.2.1.4, privateKeyPath: "~/.ssh/bfs-ed25519"} - {name: bfs-k8snode-10-2-1-5.hetzner.base.beaconfireinc.com, address: 10.2.1.5, internalAddress: 10.2.1.5, privateKeyPath: "~/.ssh/bfs-ed25519"} - {name: bfs-k8snode-10-2-1-6.hetzner.base.beaconfireinc.com, address: 10.2.1.6, internalAddress: 10.2.1.6, privateKeyPath: "~/.ssh/bfs-ed25519"} - {name: bfs-k8snode-10-2-1-7.hetzner.base.beaconfireinc.com, address: 10.2.1.7, internalAddress: 10.2.1.7, privateKeyPath: "~/.ssh/bfs-ed25519"} - {name: bfs-k8snode-10-2-1-8.hetzner.base.beaconfireinc.com, address: 10.2.1.8, internalAddress: 10.2.1.8, privateKeyPath: "~/.ssh/bfs-ed25519"} - {name: bfs-k8snode-10-2-1-9.hetzner.base.beaconfireinc.com, address: 10.2.1.9, internalAddress: 10.2.1.9, privateKeyPath: "~/.ssh/bfs-ed25519"} roleGroups: etcd: - bfs-k8smaster-10-2-1-2.hetzner.base.beaconfireinc.com - bfs-k8smaster-10-2-1-3.hetzner.base.beaconfireinc.com - bfs-k8smaster-10-2-1-4.hetzner.base.beaconfireinc.com control-plane: - bfs-k8smaster-10-2-1-2.hetzner.base.beaconfireinc.com - bfs-k8smaster-10-2-1-3.hetzner.base.beaconfireinc.com - bfs-k8smaster-10-2-1-4.hetzner.base.beaconfireinc.com worker: - bfs-k8snode-10-2-1-5.hetzner.base.beaconfireinc.com - bfs-k8snode-10-2-1-6.hetzner.base.beaconfireinc.com - bfs-k8snode-10-2-1-7.hetzner.base.beaconfireinc.com - bfs-k8snode-10-2-1-8.hetzner.base.beaconfireinc.com - bfs-k8snode-10-2-1-9.hetzner.base.beaconfireinc.com controlPlaneEndpoint: ## Internal loadbalancer for apiservers # internalLoadbalancer: haproxy domain: tcs-k8slb.beaconfireinc.com address: "10.2.1.1" port: 6443 kubernetes: version: v1.23.12 clusterName: cluster.local autoRenewCerts: true containerManager: docker etcd: type: kubekey network: plugin: calico kubePodsCIDR: 10.233.64.0/18 kubeServiceCIDR: 10.233.0.0/18 ## multus support. https://github.com/k8snetworkplumbingwg/multus-cni multusCNI: enabled: false registry: privateRegistry: "" namespaceOverride: "" registryMirrors: [] insecureRegistries: [] addons: [] --- apiVersion: installer.kubesphere.io/v1alpha1 kind: ClusterConfiguration metadata: name: ks-installer namespace: kubesphere-system labels: version: v3.3.2 spec: persistence: storageClass: "" authentication: jwtSecret: "" zone: "" local_registry: "" namespace_override: "" # dev_tag: "" etcd: monitoring: false endpointIps: localhost port: 2379 tlsEnable: true common: core: console: enableMultiLogin: true port: 30880 type: NodePort # apiserver: # resources: {} # controllerManager: # resources: {} redis: enabled: false volumeSize: 2Gi openldap: enabled: false volumeSize: 2Gi minio: volumeSize: 20Gi monitoring: # type: external endpoint: http://prometheus-operated.kubesphere-monitoring-system.svc:9090 GPUMonitoring: enabled: false gpu: kinds: - resourceName: "nvidia.com/gpu" resourceType: "GPU" default: true es: # master: # volumeSize: 4Gi # replicas: 1 # resources: {} # data: # volumeSize: 20Gi # replicas: 1 # resources: {} logMaxAge: 7 elkPrefix: logstash basicAuth: enabled: false username: "" password: "" externalElasticsearchHost: "" externalElasticsearchPort: "" alerting: enabled: false # thanosruler: # replicas: 1 # resources: {} auditing: enabled: false # operator: # resources: {} # webhook: # resources: {} devops: enabled: false # resources: {} jenkinsMemoryLim: 2Gi jenkinsMemoryReq: 500Mi jenkinsVolumeSize: 8Gi events: enabled: false # operator: # resources: {} # exporter: # resources: {} # ruler: # enabled: true # replicas: 2 # resources: {} logging: enabled: false logsidecar: enabled: true replicas: 2 # resources: {} metrics_server: enabled: false monitoring: storageClass: "" node_exporter: port: 9100 # resources: {} # kube_rbac_proxy: # resources: {} # kube_state_metrics: # resources: {} # prometheus: # replicas: 1 # volumeSize: 20Gi # resources: {} # operator: # resources: {} # alertmanager: # replicas: 1 # resources: {} # notification_manager: # resources: {} # operator: # resources: {} # proxy: # resources: {} gpu: nvidia_dcgm_exporter: enabled: false # resources: {} multicluster: clusterRole: none network: networkpolicy: enabled: false ippool: type: none topology: type: none openpitrix: store: enabled: false servicemesh: enabled: false istio: components: ingressGateways: - name: istio-ingressgateway enabled: false cni: enabled: false edgeruntime: enabled: false kubeedge: enabled: false cloudCore: cloudHub: advertiseAddress: - "" service: cloudhubNodePort: "30000" cloudhubQuicNodePort: "30001" cloudhubHttpsNodePort: "30002" cloudstreamNodePort: "30003" tunnelNodePort: "30004" # resources: {} # hostNetWork: false iptables-manager: enabled: true mode: "external" # resources: {} # edgeService: # resources: {} terminal: timeout: 600