KubeZero/charts/kubezero-addons
2023-04-12 11:14:31 +00:00
..
charts V1.24 beta, metrics updates, minor tweaks and fixes 2023-01-11 12:08:18 +00:00
templates feat: first try of 1.24 upgrade flow 2022-11-09 17:08:22 +01:00
Chart.yaml First draft of v1.25.8 2023-04-12 11:14:31 +00:00
nth.patch fix: various upgrade fixes 2022-09-20 14:54:33 +02:00
README.md First draft of v1.25.8 2023-04-12 11:14:31 +00:00
README.md.gotmpl feat: latest 1.23 cluster upgrade config 2022-09-14 19:08:14 +02:00
ruh.patch V1.24 beta, metrics updates, minor tweaks and fixes 2023-01-11 12:08:18 +00:00
update.sh V1.24 beta, metrics updates, minor tweaks and fixes 2023-01-11 12:08:18 +00:00
values.yaml First draft of v1.25.8 2023-04-12 11:14:31 +00:00

kubezero-addons

Version: 0.7.4 Type: application AppVersion: v1.24

KubeZero umbrella chart for various optional cluster addons

Homepage: https://kubezero.com

Maintainers

Name Email Url
Stefan Reimer stefan@zero-downtime.net

Requirements

Kubernetes: >= 1.24.0

Repository Name Version
aws-eks-asg-rolling-update-handler 1.2.7
aws-node-termination-handler 0.20.1
https://bitnami-labs.github.io/sealed-secrets sealed-secrets 2.7.3
https://kubernetes-sigs.github.io/external-dns/ external-dns 1.11.0
https://kubernetes.github.io/autoscaler cluster-autoscaler 9.24.0
https://nvidia.github.io/k8s-device-plugin nvidia-device-plugin 0.13.0

MetalLB

device-plugins

AWS Neuron

Device plugin for AWS Neuron - Inf1 instances

Nvidia

Cluster AutoScaler

Values

Key Type Default Description
aws-eks-asg-rolling-update-handler.enabled bool false
aws-eks-asg-rolling-update-handler.environmentVars[0].name string "CLUSTER_NAME"
aws-eks-asg-rolling-update-handler.environmentVars[0].value string ""
aws-eks-asg-rolling-update-handler.environmentVars[1].name string "AWS_REGION"
aws-eks-asg-rolling-update-handler.environmentVars[1].value string "us-west-2"
aws-eks-asg-rolling-update-handler.environmentVars[2].name string "EXECUTION_INTERVAL"
aws-eks-asg-rolling-update-handler.environmentVars[2].value string "60"
aws-eks-asg-rolling-update-handler.environmentVars[3].name string "METRICS"
aws-eks-asg-rolling-update-handler.environmentVars[3].value string "true"
aws-eks-asg-rolling-update-handler.environmentVars[4].name string "EAGER_CORDONING"
aws-eks-asg-rolling-update-handler.environmentVars[4].value string "true"
aws-eks-asg-rolling-update-handler.environmentVars[5].name string "SLOW_MODE"
aws-eks-asg-rolling-update-handler.environmentVars[5].value string "true"
aws-eks-asg-rolling-update-handler.environmentVars[6].name string "AWS_ROLE_ARN"
aws-eks-asg-rolling-update-handler.environmentVars[6].value string ""
aws-eks-asg-rolling-update-handler.environmentVars[7].name string "AWS_WEB_IDENTITY_TOKEN_FILE"
aws-eks-asg-rolling-update-handler.environmentVars[7].value string "/var/run/secrets/sts.amazonaws.com/serviceaccount/token"
aws-eks-asg-rolling-update-handler.environmentVars[8].name string "AWS_STS_REGIONAL_ENDPOINTS"
aws-eks-asg-rolling-update-handler.environmentVars[8].value string "regional"
aws-eks-asg-rolling-update-handler.image.tag string "v1.7.0"
aws-eks-asg-rolling-update-handler.nodeSelector."node-role.kubernetes.io/control-plane" string ""
aws-eks-asg-rolling-update-handler.resources.limits.memory string "128Mi"
aws-eks-asg-rolling-update-handler.resources.requests.cpu string "10m"
aws-eks-asg-rolling-update-handler.resources.requests.memory string "32Mi"
aws-eks-asg-rolling-update-handler.tolerations[0].effect string "NoSchedule"
aws-eks-asg-rolling-update-handler.tolerations[0].key string "node-role.kubernetes.io/master"
aws-eks-asg-rolling-update-handler.tolerations[1].effect string "NoSchedule"
aws-eks-asg-rolling-update-handler.tolerations[1].key string "node-role.kubernetes.io/control-plane"
aws-node-termination-handler.deleteLocalData bool true
aws-node-termination-handler.emitKubernetesEvents bool true
aws-node-termination-handler.enableProbesServer bool true
aws-node-termination-handler.enablePrometheusServer bool false
aws-node-termination-handler.enableSpotInterruptionDraining bool false
aws-node-termination-handler.enableSqsTerminationDraining bool true
aws-node-termination-handler.enabled bool false
aws-node-termination-handler.extraEnv[0] object {"name":"AWS_ROLE_ARN","value":""} "arn:aws:iam::${AWS::AccountId}:role/${AWS::Region}.${ClusterName}.awsNth"
aws-node-termination-handler.extraEnv[1].name string "AWS_WEB_IDENTITY_TOKEN_FILE"
aws-node-termination-handler.extraEnv[1].value string "/var/run/secrets/sts.amazonaws.com/serviceaccount/token"
aws-node-termination-handler.extraEnv[2].name string "AWS_STS_REGIONAL_ENDPOINTS"
aws-node-termination-handler.extraEnv[2].value string "regional"
aws-node-termination-handler.fullnameOverride string "aws-node-termination-handler"
aws-node-termination-handler.ignoreDaemonSets bool true
aws-node-termination-handler.jsonLogging bool true
aws-node-termination-handler.logFormatVersion int 2
aws-node-termination-handler.managedTag string "aws-node-termination-handler/managed" "aws-node-termination-handler/${ClusterName}"
aws-node-termination-handler.metadataTries int 0
aws-node-termination-handler.nodeSelector."node-role.kubernetes.io/control-plane" string ""
aws-node-termination-handler.podMonitor.create bool false
aws-node-termination-handler.queueURL string "" https://sqs.${AWS::Region}.amazonaws.com/${AWS::AccountId}/${ClusterName}_Nth
aws-node-termination-handler.rbac.pspEnabled bool false
aws-node-termination-handler.taintNode bool true
aws-node-termination-handler.tolerations[0].effect string "NoSchedule"
aws-node-termination-handler.tolerations[0].key string "node-role.kubernetes.io/master"
aws-node-termination-handler.tolerations[1].effect string "NoSchedule"
aws-node-termination-handler.tolerations[1].key string "node-role.kubernetes.io/control-plane"
aws-node-termination-handler.useProviderId bool true
awsNeuron.enabled bool false
awsNeuron.image.name string "public.ecr.aws/neuron/neuron-device-plugin"
awsNeuron.image.tag string "1.9.3.0"
cluster-autoscaler.autoDiscovery.clusterName string ""
cluster-autoscaler.awsRegion string "us-west-2"
cluster-autoscaler.enabled bool false
cluster-autoscaler.extraArgs.scan-interval string "30s"
cluster-autoscaler.extraArgs.skip-nodes-with-local-storage bool false
cluster-autoscaler.image.tag string "v1.24.0"
cluster-autoscaler.nodeSelector."node-role.kubernetes.io/control-plane" string ""
cluster-autoscaler.podDisruptionBudget bool false
cluster-autoscaler.prometheusRule.enabled bool false
cluster-autoscaler.prometheusRule.interval string "30"
cluster-autoscaler.serviceMonitor.enabled bool false
cluster-autoscaler.serviceMonitor.interval string "30s"
cluster-autoscaler.tolerations[0].effect string "NoSchedule"
cluster-autoscaler.tolerations[0].key string "node-role.kubernetes.io/master"
cluster-autoscaler.tolerations[1].effect string "NoSchedule"
cluster-autoscaler.tolerations[1].key string "node-role.kubernetes.io/control-plane"
clusterBackup.enabled bool false
clusterBackup.extraEnv list []
clusterBackup.image.name string "public.ecr.aws/zero-downtime/kubezero-admin"
clusterBackup.password string "" /etc/cloudbender/clusterBackup.passphrase
clusterBackup.repository string "" s3:https://s3.amazonaws.com/${CFN[ConfigBucket]}/k8s/${CLUSTERNAME}/clusterBackup
external-dns.enabled bool false
external-dns.interval string "3m"
external-dns.nodeSelector."node-role.kubernetes.io/control-plane" string ""
external-dns.provider string "inmemory"
external-dns.sources[0] string "service"
external-dns.tolerations[0].effect string "NoSchedule"
external-dns.tolerations[0].key string "node-role.kubernetes.io/master"
external-dns.tolerations[1].effect string "NoSchedule"
external-dns.tolerations[1].key string "node-role.kubernetes.io/control-plane"
external-dns.triggerLoopOnEvent bool true
forseti.aws.iamRoleArn string "" "arn:aws:iam::${AWS::AccountId}:role/${AWS::Region}.${ClusterName}.kubezeroForseti"
forseti.aws.region string ""
forseti.enabled bool false
forseti.image.name string "public.ecr.aws/zero-downtime/forseti"
forseti.image.tag string "v0.1.2"
fuseDevicePlugin.enabled bool false
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].key string "node.kubernetes.io/instance-type"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].operator string "In"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[0] string "g5.xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[10] string "g4dn.4xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[11] string "g4dn.8xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[12] string "g4dn.12xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[13] string "g4dn.16xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[1] string "g5.2xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[2] string "g5.4xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[3] string "g5.8xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[4] string "g5.12xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[5] string "g5.16xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[6] string "g5.24xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[7] string "g5.48xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[8] string "g4dn.xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[9] string "g4dn.2xlarge"
nvidia-device-plugin.enabled bool false
nvidia-device-plugin.tolerations[0].effect string "NoSchedule"
nvidia-device-plugin.tolerations[0].key string "nvidia.com/gpu"
nvidia-device-plugin.tolerations[0].operator string "Exists"
nvidia-device-plugin.tolerations[1].effect string "NoSchedule"
nvidia-device-plugin.tolerations[1].key string "kubezero-workergroup"
nvidia-device-plugin.tolerations[1].operator string "Exists"
sealed-secrets.enabled bool false
sealed-secrets.fullnameOverride string "sealed-secrets-controller"
sealed-secrets.keyrenewperiod string "0"
sealed-secrets.metrics.serviceMonitor.enabled bool false
sealed-secrets.nodeSelector."node-role.kubernetes.io/control-plane" string ""
sealed-secrets.resources.limits.memory string "128Mi"
sealed-secrets.resources.requests.cpu string "10m"
sealed-secrets.resources.requests.memory string "24Mi"
sealed-secrets.tolerations[0].effect string "NoSchedule"
sealed-secrets.tolerations[0].key string "node-role.kubernetes.io/master"
sealed-secrets.tolerations[1].effect string "NoSchedule"
sealed-secrets.tolerations[1].key string "node-role.kubernetes.io/control-plane"