KubeZero/charts/kubezero-addons
2024-12-20 03:02:59 +00:00
..
charts feat: first working v1.30 base 2024-10-16 12:20:25 +01:00
templates feat: nvidia runtimeClass, various cleanups for 1.30 2024-11-08 16:45:10 +00:00
aws-eks-asg-rolling-update-handler.patch Migrate addons to new update-lib 2023-08-20 15:38:51 +00:00
aws-node-termination-handler.patch Migrate addons to new update-lib 2023-08-20 15:38:51 +00:00
Chart.yaml chore(deps): update kubezero-addons-dependencies 2024-12-20 03:02:59 +00:00
README.md feat: add py-kube-downscaler to addons module 2024-11-13 18:19:14 +00:00
README.md.gotmpl feat: latest 1.23 cluster upgrade config 2022-09-14 19:08:14 +02:00
update.sh feat: first working v1.30 base 2024-10-16 12:20:25 +01:00
values.yaml fix: various tweaks and fixes 2024-12-04 18:40:10 +00:00

kubezero-addons

Version: 0.8.11 Type: application AppVersion: v1.30

KubeZero umbrella chart for various optional cluster addons

Homepage: https://kubezero.com

Maintainers

Name Email Url
Stefan Reimer stefan@zero-downtime.net

Requirements

Kubernetes: >= 1.26.0

Repository Name Version
https://bitnami-labs.github.io/sealed-secrets sealed-secrets 2.16.2
https://caas-team.github.io/helm-charts/ py-kube-downscaler 0.2.11
https://kubernetes-sigs.github.io/external-dns/ external-dns 1.15.0
https://kubernetes.github.io/autoscaler cluster-autoscaler 9.43.2
https://nvidia.github.io/k8s-device-plugin nvidia-device-plugin 0.17.0
https://twin.github.io/helm-charts aws-eks-asg-rolling-update-handler 1.5.0
oci://public.ecr.aws/aws-ec2/helm aws-node-termination-handler 0.24.1

MetalLB

device-plugins

AWS Neuron

Device plugin for AWS Neuron - Inf1 instances

Nvidia

Cluster AutoScaler

Values

Key Type Default Description
aws-eks-asg-rolling-update-handler.containerSecurityContext.allowPrivilegeEscalation bool false
aws-eks-asg-rolling-update-handler.containerSecurityContext.capabilities.drop[0] string "ALL"
aws-eks-asg-rolling-update-handler.enabled bool false
aws-eks-asg-rolling-update-handler.environmentVars[0].name string "CLUSTER_NAME"
aws-eks-asg-rolling-update-handler.environmentVars[0].value string ""
aws-eks-asg-rolling-update-handler.environmentVars[1].name string "AWS_REGION"
aws-eks-asg-rolling-update-handler.environmentVars[1].value string "us-west-2"
aws-eks-asg-rolling-update-handler.environmentVars[2].name string "EXECUTION_INTERVAL"
aws-eks-asg-rolling-update-handler.environmentVars[2].value string "60"
aws-eks-asg-rolling-update-handler.environmentVars[3].name string "METRICS"
aws-eks-asg-rolling-update-handler.environmentVars[3].value string "true"
aws-eks-asg-rolling-update-handler.environmentVars[4].name string "EAGER_CORDONING"
aws-eks-asg-rolling-update-handler.environmentVars[4].value string "true"
aws-eks-asg-rolling-update-handler.environmentVars[5].name string "SLOW_MODE"
aws-eks-asg-rolling-update-handler.environmentVars[5].value string "true"
aws-eks-asg-rolling-update-handler.environmentVars[6].name string "AWS_ROLE_ARN"
aws-eks-asg-rolling-update-handler.environmentVars[6].value string ""
aws-eks-asg-rolling-update-handler.environmentVars[7].name string "AWS_WEB_IDENTITY_TOKEN_FILE"
aws-eks-asg-rolling-update-handler.environmentVars[7].value string "/var/run/secrets/sts.amazonaws.com/serviceaccount/token"
aws-eks-asg-rolling-update-handler.environmentVars[8].name string "AWS_STS_REGIONAL_ENDPOINTS"
aws-eks-asg-rolling-update-handler.environmentVars[8].value string "regional"
aws-eks-asg-rolling-update-handler.image.repository string "twinproduction/aws-eks-asg-rolling-update-handler"
aws-eks-asg-rolling-update-handler.image.tag string "v1.8.4"
aws-eks-asg-rolling-update-handler.nodeSelector."node-role.kubernetes.io/control-plane" string ""
aws-eks-asg-rolling-update-handler.resources.limits.memory string "128Mi"
aws-eks-asg-rolling-update-handler.resources.requests.cpu string "10m"
aws-eks-asg-rolling-update-handler.resources.requests.memory string "32Mi"
aws-eks-asg-rolling-update-handler.securityContext.runAsNonRoot bool true
aws-eks-asg-rolling-update-handler.securityContext.runAsUser int 1001
aws-eks-asg-rolling-update-handler.securityContext.seccompProfile.type string "RuntimeDefault"
aws-eks-asg-rolling-update-handler.tolerations[0].effect string "NoSchedule"
aws-eks-asg-rolling-update-handler.tolerations[0].key string "node-role.kubernetes.io/control-plane"
aws-node-termination-handler.deleteLocalData bool true
aws-node-termination-handler.emitKubernetesEvents bool true
aws-node-termination-handler.enableProbesServer bool true
aws-node-termination-handler.enablePrometheusServer bool false
aws-node-termination-handler.enableSpotInterruptionDraining bool false
aws-node-termination-handler.enableSqsTerminationDraining bool true
aws-node-termination-handler.enabled bool false
aws-node-termination-handler.extraEnv[0] object {"name":"AWS_ROLE_ARN","value":""} "arn:aws:iam::${AWS::AccountId}:role/${AWS::Region}.${ClusterName}.awsNth"
aws-node-termination-handler.extraEnv[1].name string "AWS_WEB_IDENTITY_TOKEN_FILE"
aws-node-termination-handler.extraEnv[1].value string "/var/run/secrets/sts.amazonaws.com/serviceaccount/token"
aws-node-termination-handler.extraEnv[2].name string "AWS_STS_REGIONAL_ENDPOINTS"
aws-node-termination-handler.extraEnv[2].value string "regional"
aws-node-termination-handler.fullnameOverride string "aws-node-termination-handler"
aws-node-termination-handler.ignoreDaemonSets bool true
aws-node-termination-handler.jsonLogging bool true
aws-node-termination-handler.logFormatVersion int 2
aws-node-termination-handler.managedTag string "zdt:kubezero:nth:${ClusterName}" "zdt:kubezero:nth:${ClusterName}"
aws-node-termination-handler.metadataTries int 0
aws-node-termination-handler.nodeSelector."node-role.kubernetes.io/control-plane" string ""
aws-node-termination-handler.podMonitor.create bool false
aws-node-termination-handler.queueURL string "" https://sqs.${AWS::Region}.amazonaws.com/${AWS::AccountId}/${ClusterName}_Nth
aws-node-termination-handler.rbac.pspEnabled bool false
aws-node-termination-handler.taintNode bool true
aws-node-termination-handler.tolerations[0].effect string "NoSchedule"
aws-node-termination-handler.tolerations[0].key string "node-role.kubernetes.io/control-plane"
aws-node-termination-handler.useProviderId bool true
awsNeuron.enabled bool false
awsNeuron.image.name string "public.ecr.aws/neuron/neuron-device-plugin"
awsNeuron.image.tag string "2.22.4.0"
cluster-autoscaler.autoDiscovery.clusterName string ""
cluster-autoscaler.awsRegion string "us-west-2"
cluster-autoscaler.enabled bool false
cluster-autoscaler.extraArgs.balance-similar-node-groups bool true
cluster-autoscaler.extraArgs.ignore-taint string "node.cilium.io/agent-not-ready"
cluster-autoscaler.extraArgs.scan-interval string "30s"
cluster-autoscaler.extraArgs.skip-nodes-with-local-storage bool false
cluster-autoscaler.image.repository string "registry.k8s.io/autoscaling/cluster-autoscaler"
cluster-autoscaler.image.tag string "v1.30.2"
cluster-autoscaler.nodeSelector."node-role.kubernetes.io/control-plane" string ""
cluster-autoscaler.podDisruptionBudget bool false
cluster-autoscaler.prometheusRule.enabled bool false
cluster-autoscaler.prometheusRule.interval string "30"
cluster-autoscaler.serviceMonitor.enabled bool false
cluster-autoscaler.serviceMonitor.interval string "30s"
cluster-autoscaler.tolerations[0].effect string "NoSchedule"
cluster-autoscaler.tolerations[0].key string "node-role.kubernetes.io/control-plane"
clusterBackup.enabled bool false
clusterBackup.extraEnv list []
clusterBackup.image.name string "public.ecr.aws/zero-downtime/kubezero-admin"
clusterBackup.password string "" /etc/cloudbender/clusterBackup.passphrase
clusterBackup.repository string "" s3:https://s3.amazonaws.com/${CFN[ConfigBucket]}/k8s/${CLUSTERNAME}/clusterBackup
external-dns.enabled bool false
external-dns.interval string "3m"
external-dns.nodeSelector."node-role.kubernetes.io/control-plane" string ""
external-dns.provider string "inmemory"
external-dns.sources[0] string "service"
external-dns.tolerations[0].effect string "NoSchedule"
external-dns.tolerations[0].key string "node-role.kubernetes.io/control-plane"
external-dns.triggerLoopOnEvent bool true
forseti.aws.iamRoleArn string "" "arn:aws:iam::${AWS::AccountId}:role/${AWS::Region}.${ClusterName}.kubezeroForseti"
forseti.aws.region string ""
forseti.enabled bool false
forseti.image.name string "public.ecr.aws/zero-downtime/forseti"
forseti.image.tag string "v0.1.2"
fuseDevicePlugin.enabled bool false
fuseDevicePlugin.image.name string "public.ecr.aws/zero-downtime/fuse-device-plugin"
fuseDevicePlugin.image.tag string "v1.2.0"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].key string "node.kubernetes.io/instance-type"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].operator string "In"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[0] string "g5.xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[10] string "g4dn.4xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[11] string "g4dn.8xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[12] string "g4dn.12xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[13] string "g4dn.16xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[1] string "g5.2xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[2] string "g5.4xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[3] string "g5.8xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[4] string "g5.12xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[5] string "g5.16xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[6] string "g5.24xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[7] string "g5.48xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[8] string "g4dn.xlarge"
nvidia-device-plugin.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms[0].matchExpressions[0].values[9] string "g4dn.2xlarge"
nvidia-device-plugin.cdi.nvidiaHookPath string "/usr/bin"
nvidia-device-plugin.deviceDiscoveryStrategy string "nvml"
nvidia-device-plugin.enabled bool false
nvidia-device-plugin.runtimeClassName string "nvidia"
nvidia-device-plugin.tolerations[0].effect string "NoSchedule"
nvidia-device-plugin.tolerations[0].key string "nvidia.com/gpu"
nvidia-device-plugin.tolerations[0].operator string "Exists"
nvidia-device-plugin.tolerations[1].effect string "NoSchedule"
nvidia-device-plugin.tolerations[1].key string "kubezero-workergroup"
nvidia-device-plugin.tolerations[1].operator string "Exists"
sealed-secrets.enabled bool false
sealed-secrets.fullnameOverride string "sealed-secrets-controller"
sealed-secrets.keyrenewperiod string "0"
sealed-secrets.metrics.serviceMonitor.enabled bool false
sealed-secrets.nodeSelector."node-role.kubernetes.io/control-plane" string ""
sealed-secrets.resources.limits.memory string "128Mi"
sealed-secrets.resources.requests.cpu string "10m"
sealed-secrets.resources.requests.memory string "24Mi"
sealed-secrets.tolerations[0].effect string "NoSchedule"
sealed-secrets.tolerations[0].key string "node-role.kubernetes.io/control-plane"