From 1171d13e86dece0f85d0d972cd56c530ba883290 Mon Sep 17 00:00:00 2001 From: Stefan Reimer Date: Wed, 9 Nov 2022 17:08:22 +0100 Subject: [PATCH] feat: first try of 1.24 upgrade flow --- admin/kubezero.sh | 2 +- admin/migrate_argo_values.py | 28 ---------- admin/upgrade_cluster.sh | 40 ++++---------- admin/v1.23/pre-upgrade.sh | 52 ------------------- admin/v1.24/pre-upgrade.sh | 21 ++++++++ .../templates/cluster-backup/cronjob.yaml | 1 + charts/kubezero-ci/README.md | 16 +++--- charts/kubezero-mq/README.md | 10 ++-- charts/kubezero-network/Chart.yaml | 8 +-- charts/kubezero-network/README.md | 18 +++---- charts/kubezero-network/values.yaml | 20 +++---- charts/kubezero-sql/README.md | 3 ++ charts/kubezero-sql/values.yaml | 7 +++ charts/kubezero/README.md | 2 +- charts/kubezero/templates/network.yaml | 10 ---- charts/kubezero/values.yaml | 2 +- charts/uptime-kuma/README.md | 10 ++-- 17 files changed, 82 insertions(+), 168 deletions(-) delete mode 100644 admin/v1.23/pre-upgrade.sh create mode 100644 admin/v1.24/pre-upgrade.sh diff --git a/admin/kubezero.sh b/admin/kubezero.sh index 9e40051..cb4b60c 100755 --- a/admin/kubezero.sh +++ b/admin/kubezero.sh @@ -56,7 +56,7 @@ render_kubeadm() { cat ${WORKDIR}/kubeadm/templates/${f}Configuration.yaml >> ${HOSTFS}/etc/kubernetes/kubeadm.yaml done - # hack to "uncloack" the json patches after they go processed by helm + # "uncloak" the json patches after they got processed by helm for s in apiserver controller-manager scheduler; do yq eval '.json' ${WORKDIR}/kubeadm/templates/patches/kube-${s}1\+json.yaml > /tmp/_tmp.yaml && \ mv /tmp/_tmp.yaml ${WORKDIR}/kubeadm/templates/patches/kube-${s}1\+json.yaml diff --git a/admin/migrate_argo_values.py b/admin/migrate_argo_values.py index 35ed68d..b5c2791 100755 --- a/admin/migrate_argo_values.py +++ b/admin/migrate_argo_values.py @@ -8,34 +8,6 @@ import yaml def migrate(values): """Actual changes here""" - # migrate ClusterName to clusterName - if "ClusterName" in values: - values["clusterName"] = values["ClusterName"] - values.pop("ClusterName") - - # Remove HighAvailableControlplane - try: - values["global"]["highAvailable"] = values["HighAvailableControlplane"] - values.pop("HighAvailableControlplane") - except KeyError: - pass - - # Create new clusterwide cloudprovider data if possible - # IamArn: arn:aws:iam:::role/..cert-manager - try: - if values["cert-manager"]["IamArn"]: - account_id = values["cert-manager"]["IamArn"].split(":")[4] - region = values["cert-manager"]["IamArn"].split(":")[5].split('.')[0].split('/')[1] - if "global" not in values: - values["global"] = {} - if "aws" not in values["global"]: - values["global"]["aws"] = {} - - values["global"]["aws"]["region"] = region - values["global"]["aws"]["accountId"] = account_id - - except KeyError: - pass return values diff --git a/admin/upgrade_cluster.sh b/admin/upgrade_cluster.sh index ecec0c2..e2c0042 100755 --- a/admin/upgrade_cluster.sh +++ b/admin/upgrade_cluster.sh @@ -1,11 +1,12 @@ #!/bin/bash -e -VERSION="v1.23" #VERSION="latest" +VERSION="v1.24" ARGO_APP=${1:-/tmp/new-kubezero-argoapp.yaml} SCRIPT_DIR=$( cd -- "$( dirname -- "${BASH_SOURCE[0]}" )" &> /dev/null && pwd ) -. $SCRIPT_DIR/libhelm.sh +# shellcheck disable=SC1091 +. "$SCRIPT_DIR"/libhelm.sh [ -n "$DEBUG" ] && set -x @@ -38,6 +39,9 @@ spec: - key: node-role.kubernetes.io/master operator: Exists effect: NoSchedule + - key: node-role.kubernetes.io/control-plane + operator: Exists + effect: NoSchedule initContainers: - name: node-upgrade image: busybox @@ -75,7 +79,7 @@ EOF control_plane_upgrade() { TASKS="$1" - echo "Deploy cluster admin task: $TASK" + echo "Deploy cluster admin task: $TASKS" cat < $ARGO_APP @@ -186,6 +168,6 @@ echo "Please commit $ARGO_APP as the updated kubezero/application.yaml for your echo "Then head over to ArgoCD for this cluster and sync all KubeZero modules to apply remaining upgrades." echo " to continue and re-enable ArgoCD:" -read +read -r argo_used && enable_argo diff --git a/admin/v1.23/pre-upgrade.sh b/admin/v1.23/pre-upgrade.sh deleted file mode 100644 index d21b815..0000000 --- a/admin/v1.23/pre-upgrade.sh +++ /dev/null @@ -1,52 +0,0 @@ -#!/bin/bash - -# Migrate addons and network values from local kubeadm-values.yaml on controllers into CM -# - enable cilium - -# Create emtpy CM if not exists yet -kubectl get cm -n kube-system kubezero-values || \ -kubectl create configmap -n kube-system kubezero-values - -kubectl get cm -n kube-system kubeadm-values || \ -kubectl create configmap -n kube-system kubeadm-values - -# tweak local kubeadm for upgrade later on -yq eval -i '.global.clusterName = strenv(CLUSTERNAME) | - .global.highAvailable = env(HIGHAVAILABLE)' \ - ${HOSTFS}/etc/kubernetes/kubeadm-values.yaml - -# extract network -yq e '.network | - .cilium.enabled = true | - .calico.enabled = true | - .multus.enabled = true | - .multus.defaultNetworks = ["cilium"] | - .cilium.cluster.name = strenv(CLUSTERNAME) | - {"network": .}' ${HOSTFS}/etc/kubernetes/kubeadm-values.yaml > $WORKDIR/network-values.yaml - -# get current argo cd values -kubectl get application kubezero -n argocd -o yaml | yq '.spec.source.helm.values' > ${WORKDIR}/argo-values.yaml - -# merge all into new CM and set new minimal addons -yq ea '. as $item ireduce ({}; . * $item ) | - .global.clusterName = strenv(CLUSTERNAME) | - .global.highAvailable = env(HIGHAVAILABLE) | - .addons.clusterBackup.image.tag = "v1.23" ' ${WORKDIR}/network-values.yaml $WORKDIR/argo-values.yaml > $WORKDIR/kubezero-pre-values.yaml - -# tumble new config through migrate.py -cat $WORKDIR/kubezero-pre-values.yaml | migrate_argo_values.py > $WORKDIR/kubezero-values.yaml - -# Update kubezero-values CM -kubectl get cm -n kube-system kubezero-values -o=yaml | \ - yq e '.data."values.yaml" |= load_str("/tmp/kubezero/kubezero-values.yaml")' | \ - kubectl replace -f - - - -# update argo app -kubectl get application kubezero -n argocd -o yaml | \ - kubezero_chart_version=$(yq .version /charts/kubezero/Chart.yaml) \ - yq '.spec.source.helm.values |= load_str("/tmp/kubezero/kubezero-values.yaml") | .spec.source.targetRevision = strenv(kubezero_chart_version)' | \ - kubectl apply -f - - -# finally remove annotation to allow argo to sync again -kubectl patch app kubezero -n argocd --type json -p='[{"op": "remove", "path": "/metadata/annotations"}]' diff --git a/admin/v1.24/pre-upgrade.sh b/admin/v1.24/pre-upgrade.sh new file mode 100644 index 0000000..e6f24fa --- /dev/null +++ b/admin/v1.24/pre-upgrade.sh @@ -0,0 +1,21 @@ +#!/bin/bash + +# get current argo cd values +kubectl get application kubezero -n argocd -o yaml | yq '.spec.source.helm.values' > "${WORKDIR}"/argo-values.yaml + +# tumble new config through migrate.py +migrate_argo_values.py < "$WORKDIR"/argo-values.yaml > "$WORKDIR"/kubezero-values.yaml + +# Update kubezero-values CM +kubectl get cm -n kube-system kubezero-values -o=yaml | \ + yq e '.data."values.yaml" |= load_str("/tmp/kubezero/kubezero-values.yaml")' | \ + kubectl replace -f - + +# update argo app +kubectl get application kubezero -n argocd -o yaml | \ + kubezero_chart_version=$(yq .version /charts/kubezero/Chart.yaml) \ + yq '.spec.source.helm.values |= load_str("/tmp/kubezero/kubezero-values.yaml") | .spec.source.targetRevision = strenv(kubezero_chart_version)' | \ + kubectl apply -f - + +# finally remove annotation to allow argo to sync again +kubectl patch app kubezero -n argocd --type json -p='[{"op": "remove", "path": "/metadata/annotations"}]' diff --git a/charts/kubezero-addons/templates/cluster-backup/cronjob.yaml b/charts/kubezero-addons/templates/cluster-backup/cronjob.yaml index 976bc66..12c31b4 100644 --- a/charts/kubezero-addons/templates/cluster-backup/cronjob.yaml +++ b/charts/kubezero-addons/templates/cluster-backup/cronjob.yaml @@ -6,6 +6,7 @@ metadata: namespace: kube-system spec: schedule: "0 * * * *" + concurrencyPolicy: "Replace" jobTemplate: spec: backoffLimit: 1 diff --git a/charts/kubezero-ci/README.md b/charts/kubezero-ci/README.md index d2c7e87..13d8c25 100644 --- a/charts/kubezero-ci/README.md +++ b/charts/kubezero-ci/README.md @@ -1,6 +1,6 @@ # kubezero-ci -![Version: 0.5.15](https://img.shields.io/badge/Version-0.5.15-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) +![Version: 0.5.17](https://img.shields.io/badge/Version-0.5.17-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) KubeZero umbrella chart for all things CI @@ -20,7 +20,7 @@ Kubernetes: `>= 1.20.0` |------------|------|---------| | https://aquasecurity.github.io/helm-charts/ | trivy | 0.4.17 | | https://cdn.zero-downtime.net/charts/ | kubezero-lib | >= 0.1.5 | -| https://charts.jenkins.io | jenkins | 4.2.8 | +| https://charts.jenkins.io | jenkins | 4.2.10 | | https://dl.gitea.io/charts/ | gitea | 5.0.9 | | https://gocd.github.io/helm-chart | gocd | 1.40.8 | @@ -95,16 +95,16 @@ Kubernetes: `>= 1.20.0` | jenkins.controller.initContainerResources.limits.memory | string | `"1024Mi"` | | | jenkins.controller.initContainerResources.requests.cpu | string | `"50m"` | | | jenkins.controller.initContainerResources.requests.memory | string | `"256Mi"` | | -| jenkins.controller.installPlugins[0] | string | `"kubernetes:3724.v0920c1e0ec69"` | | +| jenkins.controller.installPlugins[0] | string | `"kubernetes:3734.v562b_b_a_627ea_c"` | | | jenkins.controller.installPlugins[1] | string | `"workflow-aggregator:581.v0c46fa_697ffd"` | | -| jenkins.controller.installPlugins[2] | string | `"git:4.12.1"` | | -| jenkins.controller.installPlugins[3] | string | `"configuration-as-code:1512.vb_79d418d5fc8"` | | +| jenkins.controller.installPlugins[2] | string | `"git:4.13.0"` | | +| jenkins.controller.installPlugins[3] | string | `"configuration-as-code:1569.vb_72405b_80249"` | | | jenkins.controller.installPlugins[4] | string | `"antisamy-markup-formatter:2.7"` | | | jenkins.controller.installPlugins[5] | string | `"prometheus:2.0.11"` | | | jenkins.controller.installPlugins[6] | string | `"htmlpublisher:1.31"` | | | jenkins.controller.installPlugins[7] | string | `"build-discarder:139.v05696a_7fe240"` | | -| jenkins.controller.installPlugins[8] | string | `"dark-theme:245.vb_a_2b_b_010ea_96"` | | -| jenkins.controller.installPlugins[9] | string | `"kubernetes-credentials-provider:1.199.v4a_1d1f5d074f"` | | +| jenkins.controller.installPlugins[8] | string | `"dark-theme:262.v0202a_4c8fb_6a"` | | +| jenkins.controller.installPlugins[9] | string | `"kubernetes-credentials-provider:1.206.v7ce2cf7b_0c8b"` | | | jenkins.controller.javaOpts | string | `"-XX:+UseContainerSupport -XX:+UseStringDeduplication -Dhudson.model.DirectoryBrowserSupport.CSP=\"sandbox allow-popups; default-src 'none'; img-src 'self' cdn.zero-downtime.net; style-src 'unsafe-inline';\""` | | | jenkins.controller.jenkinsOpts | string | `"--sessionTimeout=180 --sessionEviction=3600"` | | | jenkins.controller.prometheus.enabled | bool | `false` | | @@ -112,7 +112,7 @@ Kubernetes: `>= 1.20.0` | jenkins.controller.resources.limits.memory | string | `"4096Mi"` | | | jenkins.controller.resources.requests.cpu | string | `"250m"` | | | jenkins.controller.resources.requests.memory | string | `"1280Mi"` | | -| jenkins.controller.tag | string | `"2.370-alpine-jdk17"` | | +| jenkins.controller.tag | string | `"alpine-jdk17"` | | | jenkins.controller.testEnabled | bool | `false` | | | jenkins.enabled | bool | `false` | | | jenkins.istio.agent.enabled | bool | `false` | | diff --git a/charts/kubezero-mq/README.md b/charts/kubezero-mq/README.md index ee22ba3..3f0b17c 100644 --- a/charts/kubezero-mq/README.md +++ b/charts/kubezero-mq/README.md @@ -1,6 +1,6 @@ # kubezero-mq -![Version: 0.2.3](https://img.shields.io/badge/Version-0.2.3-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) +![Version: 0.3.0](https://img.shields.io/badge/Version-0.3.0-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) KubeZero umbrella chart for MQ systems like NATS, RabbitMQ @@ -19,8 +19,8 @@ Kubernetes: `>= 1.20.0` | Repository | Name | Version | |------------|------|---------| | | nats | 0.8.4 | -| https://cdn.zero-downtime.net/charts/ | kubezero-lib | >= 0.1.3 | -| https://charts.bitnami.com/bitnami | rabbitmq | 9.0.3 | +| https://cdn.zero-downtime.net/charts/ | kubezero-lib | >= 0.1.5 | +| https://charts.bitnami.com/bitnami | rabbitmq | 11.1.1 | ## Values @@ -34,8 +34,8 @@ Kubernetes: `>= 1.20.0` | nats.nats.advertise | bool | `false` | | | nats.nats.jetstream.enabled | bool | `true` | | | nats.natsbox.enabled | bool | `false` | | -| rabbitmq.auth.erlangCookie | string | `"randomlongerlangcookie"` | | -| rabbitmq.auth.password | string | `"supersecret"` | | +| rabbitmq.auth.existingErlangSecret | string | `"rabbitmq"` | | +| rabbitmq.auth.existingPasswordSecret | string | `"rabbitmq"` | | | rabbitmq.auth.tls.enabled | bool | `false` | | | rabbitmq.auth.tls.existingSecret | string | `"rabbitmq-server-certificate"` | | | rabbitmq.auth.tls.existingSecretFullChain | bool | `true` | | diff --git a/charts/kubezero-network/Chart.yaml b/charts/kubezero-network/Chart.yaml index 933c9f2..2da7050 100644 --- a/charts/kubezero-network/Chart.yaml +++ b/charts/kubezero-network/Chart.yaml @@ -2,7 +2,7 @@ apiVersion: v2 name: kubezero-network description: KubeZero umbrella chart for all things network type: application -version: 0.4.0 +version: 0.4.1 home: https://kubezero.com icon: https://cdn.zero-downtime.net/assets/kubezero/logo-small-64.png keywords: @@ -19,14 +19,14 @@ dependencies: version: ">= 0.1.5" repository: https://cdn.zero-downtime.net/charts/ - name: cilium - version: 1.12.2 + version: 1.12.3 repository: https://helm.cilium.io/ condition: cilium.enabled - name: metallb - version: 0.13.5 + version: 0.13.7 repository: https://metallb.github.io/metallb condition: metallb.enabled - name: calico version: 0.2.2 condition: calico.enabled -kubeVersion: ">= 1.24.0" +kubeVersion: ">= 1.20.0" diff --git a/charts/kubezero-network/README.md b/charts/kubezero-network/README.md index 9da1a3b..4fddb95 100644 --- a/charts/kubezero-network/README.md +++ b/charts/kubezero-network/README.md @@ -1,6 +1,6 @@ # kubezero-network -![Version: 0.4.0](https://img.shields.io/badge/Version-0.4.0-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) +![Version: 0.4.1](https://img.shields.io/badge/Version-0.4.1-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) KubeZero umbrella chart for all things network @@ -14,14 +14,14 @@ KubeZero umbrella chart for all things network ## Requirements -Kubernetes: `>= 1.24.0` +Kubernetes: `>= 1.20.0` | Repository | Name | Version | |------------|------|---------| | | calico | 0.2.2 | | https://cdn.zero-downtime.net/charts/ | kubezero-lib | >= 0.1.5 | -| https://helm.cilium.io/ | cilium | 1.12.2 | -| https://metallb.github.io/metallb | metallb | 0.13.5 | +| https://helm.cilium.io/ | cilium | 1.12.3 | +| https://metallb.github.io/metallb | metallb | 0.13.7 | ## Values @@ -34,22 +34,19 @@ Kubernetes: `>= 1.24.0` | cilium.cluster.id | int | `240` | | | cilium.cluster.name | string | `"default"` | | | cilium.cni.binPath | string | `"/usr/libexec/cni"` | | -| cilium.cni.chainingMode | string | `"generic-veth"` | | -| cilium.cni.exclusive | bool | `false` | | | cilium.containerRuntime.integration | string | `"crio"` | | | cilium.enabled | bool | `false` | | | cilium.hubble.enabled | bool | `false` | | | cilium.ipam.operator.clusterPoolIPv4PodCIDRList[0] | string | `"10.240.0.0/16"` | | | cilium.l2NeighDiscovery.enabled | bool | `false` | | | cilium.l7Proxy | bool | `false` | | -| cilium.nodePort.enabled | bool | `false` | | +| cilium.nodePort.enabled | bool | `true` | | | cilium.operator.nodeSelector."node-role.kubernetes.io/control-plane" | string | `""` | | | cilium.operator.replicas | int | `1` | | | cilium.operator.tolerations[0].effect | string | `"NoSchedule"` | | | cilium.operator.tolerations[0].key | string | `"node-role.kubernetes.io/master"` | | | cilium.operator.tolerations[1].effect | string | `"NoSchedule"` | | | cilium.operator.tolerations[1].key | string | `"node-role.kubernetes.io/control-plane"` | | -| cilium.policyEnforcementMode | string | `"never"` | | | cilium.prometheus.enabled | bool | `false` | | | cilium.prometheus.port | int | `9091` | | | cilium.securityContext.privileged | bool | `true` | | @@ -61,8 +58,5 @@ Kubernetes: `>= 1.24.0` | metallb.controller.tolerations[1].key | string | `"node-role.kubernetes.io/control-plane"` | | | metallb.enabled | bool | `false` | | | metallb.ipAddressPools | list | `[]` | | -| multus.clusterNetwork | string | `"calico"` | | -| multus.defaultNetworks | list | `[]` | | | multus.enabled | bool | `false` | | -| multus.readinessindicatorfile | string | `"/etc/cni/net.d/10-calico.conflist"` | | -| multus.tag | string | `"v3.9.1"` | | +| multus.tag | string | `"v3.9.2"` | | diff --git a/charts/kubezero-network/values.yaml b/charts/kubezero-network/values.yaml index fb03552..5714e0e 100644 --- a/charts/kubezero-network/values.yaml +++ b/charts/kubezero-network/values.yaml @@ -18,12 +18,12 @@ metallb: multus: enabled: false - tag: "v3.9.1" + tag: "v3.9.2" - clusterNetwork: "calico" - defaultNetworks: [] - # - "cilium" - readinessindicatorfile: "/etc/cni/net.d/10-calico.conflist" + #clusterNetwork: "calico" + #defaultNetworks: [] + # - "cilium" + #readinessindicatorfile: "/etc/cni/net.d/10-calico.conflist" cilium: enabled: false @@ -38,17 +38,13 @@ cilium: cni: binPath: "/usr/libexec/cni" #-- Ensure this is false if multus is enabled - exclusive: false - chainingMode: generic-veth + # exclusive: false + # chainingMode: generic-veth bpf: hostLegacyRouting: true # tproxy: false - # enableIPv4Masquerade: false - # enableIdentityMark: false - policyEnforcementMode: "never" - cluster: # This should match the second octet of clusterPoolIPv4PodCIDRList # to prevent IP space overlap and easy tracking @@ -63,7 +59,7 @@ cilium: # Should be handled by multus nodePort: - enabled: false + enabled: true # Keep it simple for now l7Proxy: false diff --git a/charts/kubezero-sql/README.md b/charts/kubezero-sql/README.md index daae4cb..492e81a 100644 --- a/charts/kubezero-sql/README.md +++ b/charts/kubezero-sql/README.md @@ -40,6 +40,9 @@ Kubernetes: `>= 1.20.0` | mariadb-galera.replicaCount | int | `2` | | | pxc-operator.enabled | bool | `false` | | | pxc-operator.nodeSelector."node-role.kubernetes.io/control-plane" | string | `""` | | +| pxc-operator.resources.limits.memory | string | `"512Mi"` | | +| pxc-operator.resources.requests.cpu | string | `"50m"` | | +| pxc-operator.resources.requests.memory | string | `"32Mi"` | | | pxc-operator.tolerations[0].effect | string | `"NoSchedule"` | | | pxc-operator.tolerations[0].key | string | `"node-role.kubernetes.io/master"` | | | pxc-operator.tolerations[1].effect | string | `"NoSchedule"` | | diff --git a/charts/kubezero-sql/values.yaml b/charts/kubezero-sql/values.yaml index 6366ff6..d00b3e5 100644 --- a/charts/kubezero-sql/values.yaml +++ b/charts/kubezero-sql/values.yaml @@ -13,6 +13,13 @@ pxc-operator: nodeSelector: node-role.kubernetes.io/control-plane: "" + resources: + limits: + #cpu: 200m + memory: 512Mi + requests: + cpu: 50m + memory: 32Mi mariadb-galera: enabled: false diff --git a/charts/kubezero/README.md b/charts/kubezero/README.md index 2bc9276..b4bd099 100644 --- a/charts/kubezero/README.md +++ b/charts/kubezero/README.md @@ -68,7 +68,7 @@ Kubernetes: `>= 1.24.0` | network.cilium.cluster | object | `{}` | | | network.enabled | bool | `true` | | | network.retain | bool | `true` | | -| network.targetRevision | string | `"0.4.0"` | | +| network.targetRevision | string | `"0.4.1"` | | | storage.aws-ebs-csi-driver.enabled | bool | `false` | | | storage.aws-efs-csi-driver.enabled | bool | `false` | | | storage.enabled | bool | `false` | | diff --git a/charts/kubezero/templates/network.yaml b/charts/kubezero/templates/network.yaml index 2a258f7..8aa8e97 100644 --- a/charts/kubezero/templates/network.yaml +++ b/charts/kubezero/templates/network.yaml @@ -1,15 +1,5 @@ {{- define "network-values" }} -# until 1.24 -calico: - enabled: true - -# since 1.22 -multus: - enabled: true - defaultNetworks: - - "cilium" - # since 1.23 cilium: enabled: true diff --git a/charts/kubezero/values.yaml b/charts/kubezero/values.yaml index 93f84f0..a2b7b8d 100644 --- a/charts/kubezero/values.yaml +++ b/charts/kubezero/values.yaml @@ -25,7 +25,7 @@ addons: network: enabled: true retain: true - targetRevision: 0.4.0 + targetRevision: 0.4.1 cilium: cluster: {} diff --git a/charts/uptime-kuma/README.md b/charts/uptime-kuma/README.md index 5b9f050..67f3ff0 100644 --- a/charts/uptime-kuma/README.md +++ b/charts/uptime-kuma/README.md @@ -1,8 +1,8 @@ # uptime-kuma -![Version: 0.1.10](https://img.shields.io/badge/Version-0.1.10-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) ![AppVersion: 1.9.1](https://img.shields.io/badge/AppVersion-1.9.1-informational?style=flat-square) +![Version: 0.2.0](https://img.shields.io/badge/Version-0.2.0-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) ![AppVersion: 1.18.5](https://img.shields.io/badge/AppVersion-1.18.5-informational?style=flat-square) -Chart for deploying uptime-kuma +Chart for deploying uptime-kuma on KubeZero **Homepage:** @@ -10,15 +10,15 @@ Chart for deploying uptime-kuma | Name | Email | Url | | ---- | ------ | --- | -| Quarky9 | | | +| Stefan Reimer | | | ## Requirements -Kubernetes: `>= 1.18.0` +Kubernetes: `>= 1.20.0` | Repository | Name | Version | |------------|------|---------| -| https://cdn.zero-downtime.net/charts/ | kubezero-lib | >= 0.1.4 | +| https://cdn.zero-downtime.net/charts/ | kubezero-lib | >= 0.1.5 | ## Values