2021-11-27 13:02:23 +00:00
|
|
|
#!/bin/sh
|
|
|
|
set -e
|
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
if [ -n "$DEBUG" ]; then
|
|
|
|
set -x
|
|
|
|
LOG="--v=5"
|
|
|
|
fi
|
|
|
|
|
|
|
|
# Export vars to ease use in debug_shell etc
|
|
|
|
export WORKDIR=/tmp/kubezero
|
|
|
|
export HOSTFS=/host
|
|
|
|
export VERSION=v1.21
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
export KUBECONFIG="${HOSTFS}/root/.kube/config"
|
|
|
|
|
2021-12-15 22:19:15 +00:00
|
|
|
# etcd
|
|
|
|
export ETCDCTL_API=3
|
|
|
|
export ETCDCTL_CACERT=${HOSTFS}/etc/kubernetes/pki/etcd/ca.crt
|
|
|
|
export ETCDCTL_CERT=${HOSTFS}/etc/kubernetes/pki/apiserver-etcd-client.crt
|
|
|
|
export ETCDCTL_KEY=${HOSTFS}/etc/kubernetes/pki/apiserver-etcd-client.key
|
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
# Generic retry utility
|
|
|
|
retry() {
|
|
|
|
local tries=$1
|
|
|
|
local waitfor=$2
|
|
|
|
local timeout=$3
|
|
|
|
shift 3
|
|
|
|
while true; do
|
|
|
|
type -tf $1 >/dev/null && { timeout $timeout $@ && return; } || { $@ && return; }
|
|
|
|
let tries=$tries-1
|
|
|
|
[ $tries -eq 0 ] && return 1
|
|
|
|
sleep $waitfor
|
|
|
|
done
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2021-12-15 22:19:15 +00:00
|
|
|
_kubeadm() {
|
|
|
|
kubeadm $@ --config /etc/kubernetes/kubeadm.yaml --rootfs ${HOSTFS} $LOG
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
# Render cluster config
|
|
|
|
render_kubeadm() {
|
|
|
|
helm template /opt/kubeadm --output-dir ${WORKDIR} -f ${HOSTFS}/etc/kubernetes/kubezero.yaml
|
|
|
|
|
|
|
|
# Assemble kubeadm config
|
2021-12-15 22:19:15 +00:00
|
|
|
cat /dev/null > ${HOSTFS}/etc/kubernetes/kubeadm.yaml
|
2021-11-27 13:02:23 +00:00
|
|
|
for f in Cluster Init KubeProxy Kubelet; do
|
2021-12-15 22:19:15 +00:00
|
|
|
# echo "---" >> /etc/kubernetes/kubeadm.yaml
|
|
|
|
cat ${WORKDIR}/kubeadm/templates/${f}Configuration.yaml >> ${HOSTFS}/etc/kubernetes/kubeadm.yaml
|
2021-11-27 13:02:23 +00:00
|
|
|
done
|
|
|
|
|
|
|
|
# hack to "uncloack" the json patches after they go processed by helm
|
|
|
|
for s in apiserver controller-manager scheduler; do
|
|
|
|
yq eval '.json' ${WORKDIR}/kubeadm/templates/patches/kube-${s}1\+json.yaml > /tmp/_tmp.yaml && \
|
|
|
|
mv /tmp/_tmp.yaml ${WORKDIR}/kubeadm/templates/patches/kube-${s}1\+json.yaml
|
|
|
|
done
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
parse_kubezero() {
|
|
|
|
[ -f ${HOSTFS}/etc/kubernetes/kubezero.yaml ] || { echo "Missing /etc/kubernetes/kubezero.yaml!"; exit 1; }
|
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
export KUBE_VERSION=$(kubeadm version -o yaml | yq eval .clientVersion.gitVersion -)
|
|
|
|
export CLUSTERNAME=$(yq eval '.clusterName' ${HOSTFS}/etc/kubernetes/kubezero.yaml)
|
|
|
|
export ETCD_NODENAME=$(yq eval '.etcd.nodeName' ${HOSTFS}/etc/kubernetes/kubezero.yaml)
|
2021-11-27 13:02:23 +00:00
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
export AWS_IAM_AUTH=$(yq eval '.api.awsIamAuth.enabled' ${HOSTFS}/etc/kubernetes/kubezero.yaml)
|
|
|
|
export AWS_NTH=$(yq eval '.addons.aws-node-termination-handler.enabled' ${HOSTFS}/etc/kubernetes/kubezero.yaml)
|
2021-11-27 13:02:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
# Shared steps before calling kubeadm
|
|
|
|
pre_kubeadm() {
|
|
|
|
# update all apiserver addons first
|
|
|
|
cp -r ${WORKDIR}/kubeadm/templates/apiserver ${HOSTFS}/etc/kubernetes
|
2021-12-01 12:33:11 +00:00
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
# aws-iam-authenticator enabled ?
|
|
|
|
if [ "$AWS_IAM_AUTH" == "true" ]; then
|
|
|
|
|
|
|
|
# Initialize webhook
|
|
|
|
if [ ! -f ${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.crt ]; then
|
|
|
|
aws-iam-authenticator init -i ${CLUSTERNAME}
|
|
|
|
mv key.pem ${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.key
|
|
|
|
mv cert.pem ${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.crt
|
|
|
|
fi
|
|
|
|
|
|
|
|
# Patch the aws-iam-authenticator config with the actual cert.pem
|
|
|
|
yq eval -Mi ".clusters[0].cluster.certificate-authority-data = \"$(cat ${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.crt| base64 -w0)\"" ${HOSTFS}/etc/kubernetes/apiserver/aws-iam-authenticator.yaml
|
|
|
|
fi
|
|
|
|
|
|
|
|
# copy patches to host to make --rootfs of kubeadm work
|
|
|
|
cp -r ${WORKDIR}/kubeadm/templates/patches /host/tmp/
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
# Shared steps after calling kubeadm
|
|
|
|
post_kubeadm() {
|
|
|
|
# KubeZero resources
|
2021-12-01 12:33:11 +00:00
|
|
|
for f in ${WORKDIR}/kubeadm/templates/resources/*.yaml; do
|
|
|
|
kubectl apply -f $f $LOG
|
|
|
|
done
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
# Patch coreDNS addon, ideally we prevent kubeadm to reset coreDNS to its defaults
|
|
|
|
kubectl patch deployment coredns -n kube-system --patch-file ${WORKDIR}/kubeadm/templates/patches/coredns0.yaml $LOG
|
|
|
|
|
|
|
|
rm -rf /host/tmp/patches
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
# First parse kubezero.yaml
|
|
|
|
parse_kubezero
|
|
|
|
|
|
|
|
if [ "$1" == 'upgrade' ]; then
|
|
|
|
### PRE 1.21 specific
|
|
|
|
#####################
|
|
|
|
|
|
|
|
# Migrate aws-iam-authenticator from file certs to secret
|
|
|
|
if [ "$AWS_IAM_AUTH" == "true" ]; then
|
|
|
|
kubectl get secrets -n kube-system aws-iam-certs || \
|
|
|
|
kubectl create secret generic aws-iam-certs -n kube-system \
|
|
|
|
--from-file=key.pem=${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.key \
|
|
|
|
--from-file=cert.pem=${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.crt
|
|
|
|
fi
|
|
|
|
|
|
|
|
#####################
|
|
|
|
|
|
|
|
render_kubeadm
|
|
|
|
|
|
|
|
pre_kubeadm
|
|
|
|
|
|
|
|
# Upgrade
|
2021-12-15 22:19:15 +00:00
|
|
|
_kubeadm upgrade apply -y --experimental-patches /tmp/patches
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
post_kubeadm
|
|
|
|
|
|
|
|
# If we have a re-cert kubectl config install for root
|
|
|
|
if [ -f ${HOSTFS}/etc/kubernetes/admin.conf ]; then
|
|
|
|
cp ${HOSTFS}/etc/kubernetes/admin.conf ${HOSTFS}/root/.kube/config
|
|
|
|
fi
|
|
|
|
|
|
|
|
### POST 1.21 specific
|
|
|
|
######################
|
2021-12-01 12:33:11 +00:00
|
|
|
helm repo add kubezero https://cdn.zero-downtime.net/charts/
|
|
|
|
|
2021-12-03 21:13:40 +00:00
|
|
|
# network
|
|
|
|
yq eval '.network // ""' ${HOSTFS}/etc/kubernetes/kubezero.yaml > _values.yaml
|
|
|
|
helm template kubezero/kubezero-network --version 0.1.0 --include-crds --namespace kube-system --name-template network \
|
|
|
|
-f _values.yaml --kube-version $KUBE_VERSION | kubectl apply -f - $LOG
|
2021-12-01 12:33:11 +00:00
|
|
|
|
2021-12-03 21:13:40 +00:00
|
|
|
# addons
|
|
|
|
yq eval '.addons // ""' ${HOSTFS}/etc/kubernetes/kubezero.yaml > _values.yaml
|
|
|
|
helm template kubezero/kubezero-addons --version 0.2.2 --include-crds --namespace kube-system --name-template addons \
|
|
|
|
-f _values.yaml --kube-version $KUBE_VERSION | kubectl apply -f - $LOG
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
######################
|
|
|
|
|
2021-12-03 21:13:40 +00:00
|
|
|
# Execute cluster backup to allow new controllers to join
|
|
|
|
kubectl create job backup-cluster-now --from=cronjob/kubezero-backup -n kube-system
|
|
|
|
|
|
|
|
# That might take a while as the backup pod needs the CNIs to come online etc.
|
|
|
|
retry 10 30 40 kubectl wait --for=condition=complete job/backup-cluster-now -n kube-system && kubectl delete job backup-cluster-now -n kube-system
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
# Cleanup after kubeadm on the host
|
2021-12-03 21:13:40 +00:00
|
|
|
rm -rf ${HOSTFS}/etc/kubernetes/tmp
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
echo "Successfully upgraded cluster."
|
|
|
|
|
|
|
|
# TODO
|
|
|
|
# Send Notification currently done via CloudBender -> SNS -> Slack
|
|
|
|
# Better deploy https://github.com/opsgenie/kubernetes-event-exporter and set proper routes and labels on this Job
|
|
|
|
|
|
|
|
# Removed:
|
|
|
|
# - update oidc do we need that ?
|
|
|
|
|
2021-12-03 21:13:40 +00:00
|
|
|
elif [[ "$1" == 'node-upgrade' ]]; then
|
|
|
|
echo "Starting node upgrade ..."
|
|
|
|
|
|
|
|
echo "Migrating kubezero.yaml"
|
2021-12-17 10:57:23 +00:00
|
|
|
yq -i eval '.api.etcdServers = .api.allEtcdEndpoints | .network.multus.enabled = "true"' ${HOSTFS}/etc/kubernetes/kubezero.yaml
|
|
|
|
|
|
|
|
# AWS
|
|
|
|
if [ -f ${HOSTFS}/etc/cloudbender/clusterBackup.passphrase ]; then
|
|
|
|
if [ -f ${HOSTFS}/usr/local/sbin/backup_control_plane.sh ]; then
|
|
|
|
mv ${HOSTFS}/usr/local/sbin/backup_control_plane.sh ${HOSTFS}/usr/local/sbin/backup_control_plane.disabled
|
|
|
|
echo "Disabled old cluster backup OS cronjob"
|
|
|
|
fi
|
2021-12-03 21:13:40 +00:00
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
# enable backup and awsIamAuth & multus
|
|
|
|
yq -i eval '
|
|
|
|
.api.awsIamAuth.enabled = "true" | .api.awsIamAuth.workerNodeRole = .workerNodeRole | .api.awsIamAuth.kubeAdminRole = .kubeAdminRole
|
|
|
|
| .api.serviceAccountIssuer = .serviceAccountIssuer | .api.apiAudiences = "istio-ca,sts.amazonaws.com"
|
|
|
|
' ${HOSTFS}/etc/kubernetes/kubezero.yaml
|
2021-12-03 21:13:40 +00:00
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
export restic_repo=$(grep "export RESTIC_REPOSITORY" ${HOSTFS}/usr/local/sbin/backup_control_plane.disabled | sed -e 's/.*=//' | sed -e 's/"//g')
|
|
|
|
export restic_pw="$(cat ${HOSTFS}/etc/cloudbender/clusterBackup.passphrase)"
|
|
|
|
export REGION=$(kubectl get node $NODE_NAME -o yaml | yq eval '.metadata.labels."topology.kubernetes.io/region"' -)
|
|
|
|
|
|
|
|
if [ -n "$restic_repo" ]; then
|
|
|
|
yq -i eval '
|
|
|
|
.addons.clusterBackup.enabled = "true" | .addons.clusterBackup.repository = strenv(restic_repo) | .addons.clusterBackup.password = strenv(restic_pw)
|
2021-12-19 22:18:01 +00:00
|
|
|
| .addons.clusterBackup.image.tag = strenv(KUBE_VERSION)
|
2021-12-17 10:57:23 +00:00
|
|
|
| .addons.clusterBackup.extraEnv[0].name = "AWS_DEFAULT_REGION" | .addons.clusterBackup.extraEnv[0].value = strenv(REGION)
|
|
|
|
' ${HOSTFS}/etc/kubernetes/kubezero.yaml
|
|
|
|
fi
|
|
|
|
fi
|
2021-12-03 21:13:40 +00:00
|
|
|
|
|
|
|
echo "All done."
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
elif [[ "$1" =~ "^(bootstrap|recover|join)$" ]]; then
|
|
|
|
|
|
|
|
render_kubeadm
|
|
|
|
|
|
|
|
if [[ "$1" =~ "^(recover|join)$" ]]; then
|
|
|
|
|
|
|
|
# Recert certificates for THIS node
|
|
|
|
rm -f ${HOSTFS}/etc/kubernetes/pki/etcd/peer.* ${HOSTFS}/etc/kubernetes/pki/etcd/server.* ${HOSTFS}/etc/kubernetes/pki/apiserver.*
|
2021-12-15 22:19:15 +00:00
|
|
|
_kubeadm init phase certs etcd-server
|
|
|
|
_kubeadm init phase certs etcd-peer
|
|
|
|
_kubeadm init phase certs apiserver
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
# Restore only etcd for desaster recovery
|
|
|
|
if [[ "$1" =~ "^(recover)$" ]]; then
|
2021-12-15 22:19:15 +00:00
|
|
|
etcdctl snapshot restore ${HOSTFS}/etc/kubernetes/etcd_snapshot \
|
|
|
|
--name $ETCD_NODENAME \
|
2021-11-27 13:02:23 +00:00
|
|
|
--data-dir="${HOSTFS}/var/lib/etcd" \
|
2021-12-15 22:19:15 +00:00
|
|
|
--initial-cluster-token etcd-${CLUSTERNAME} \
|
|
|
|
--initial-advertise-peer-urls https://${ETCD_NODENAME}:2380 \
|
|
|
|
--initial-cluster $ETCD_NODENAME=https://${ETCD_NODENAME}:2380
|
2021-11-27 13:02:23 +00:00
|
|
|
fi
|
|
|
|
|
|
|
|
# Create all certs during bootstrap
|
|
|
|
else
|
2021-12-15 22:19:15 +00:00
|
|
|
_kubeadm init phase certs all
|
2021-11-27 13:02:23 +00:00
|
|
|
fi
|
|
|
|
|
|
|
|
pre_kubeadm
|
|
|
|
|
|
|
|
if [[ "$1" =~ "^(join)$" ]]; then
|
2021-12-15 22:19:15 +00:00
|
|
|
|
|
|
|
_kubeadm init phase preflight
|
|
|
|
_kubeadm init phase kubeconfig all
|
|
|
|
_kubeadm init phase kubelet-start
|
|
|
|
|
|
|
|
# first get current running etcd pods for etcdctl commands
|
|
|
|
# retry in case other nodes join / API fails / etcd leader changes etc.
|
|
|
|
while true; do
|
|
|
|
etcd_endpoints=$(kubectl get pods -n kube-system -l component=etcd -o yaml | \
|
|
|
|
yq eval '.items[].metadata.annotations."kubeadm.kubernetes.io/etcd.advertise-client-urls"' - | tr '\n' ',' | sed -e 's/,$//')
|
|
|
|
[[ $etcd_endpoints =~ ^https:// ]] && break
|
|
|
|
sleep 3
|
|
|
|
done
|
|
|
|
|
|
|
|
# is our $ETCD_NODENAME already in the etcd cluster ?
|
|
|
|
# Remove former self first
|
|
|
|
MY_ID=$(etcdctl member list --endpoints=$etcd_endpoints | grep $ETCD_NODENAME | awk '{print $1}' | sed -e 's/,$//')
|
|
|
|
[ -n "$MY_ID" ] && retry 12 5 5 etcdctl member remove $MY_ID --endpoints=$etcd_endpoints
|
|
|
|
|
|
|
|
# Announce new etcd member and capture ETCD_INITIAL_CLUSTER, retry needed in case another node joining causes temp quorum loss
|
|
|
|
ETCD_ENVS=$(retry 12 5 5 etcdctl member add $ETCD_NODENAME --peer-urls="https://${ETCD_NODENAME}:2380" --endpoints=$etcd_endpoints)
|
|
|
|
export $(echo "$ETCD_ENVS" | grep ETCD_INITIAL_CLUSTER= | sed -e 's/"//g')
|
|
|
|
|
|
|
|
# Patch kubezero.yaml and re-render to get etcd manifest patched
|
|
|
|
yq eval -i '.etcd.state = "existing"
|
|
|
|
| .etcd.initialCluster = strenv(ETCD_INITIAL_CLUSTER)
|
|
|
|
' ${HOSTFS}/etc/kubernetes/kubezero.yaml
|
|
|
|
render_kubeadm
|
|
|
|
|
|
|
|
# Generate our advanced etcd yaml
|
|
|
|
_kubeadm init phase etcd local --experimental-patches /tmp/patches
|
|
|
|
|
|
|
|
_kubeadm init phase control-plane all --experimental-patches /tmp/patches
|
|
|
|
_kubeadm init phase mark-control-plane
|
|
|
|
_kubeadm init phase kubelet-finalize all
|
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
else
|
2021-12-15 22:19:15 +00:00
|
|
|
_kubeadm init --experimental-patches /tmp/patches --skip-token-print
|
2021-11-27 13:02:23 +00:00
|
|
|
fi
|
|
|
|
|
|
|
|
cp ${HOSTFS}/etc/kubernetes/admin.conf ${HOSTFS}/root/.kube/config
|
|
|
|
|
|
|
|
# Wait for api to be online
|
|
|
|
retry 0 10 30 kubectl cluster-info --request-timeout 3
|
|
|
|
|
|
|
|
# Ensure aws-iam-authenticator secret is in place
|
|
|
|
if [ "$AWS_IAM_AUTH" == "true" ]; then
|
|
|
|
kubectl get secrets -n kube-system aws-iam-certs || \
|
|
|
|
kubectl create secret generic aws-iam-certs -n kube-system \
|
|
|
|
--from-file=key.pem=${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.key \
|
|
|
|
--from-file=cert.pem=${HOSTFS}/etc/kubernetes/pki/aws-iam-authenticator.crt
|
|
|
|
|
|
|
|
# Store aws-iam-auth admin on SSM
|
|
|
|
yq eval -M ".clusters[0].cluster.certificate-authority-data = \"$(cat ${HOSTFS}/etc/kubernetes/pki/ca.crt | base64 -w0)\"" ${WORKDIR}/kubeadm/templates/admin-aws-iam.yaml > ${HOSTFS}/etc/kubernetes/admin-aws-iam.yaml
|
|
|
|
fi
|
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
# Install some basics on bootstrap and join for 1.21 to get new modules in place
|
2021-12-03 21:13:40 +00:00
|
|
|
if [[ "$1" =~ "^(bootstrap|join|recover)$" ]]; then
|
2021-12-01 12:33:11 +00:00
|
|
|
helm repo add kubezero https://cdn.zero-downtime.net/charts/
|
|
|
|
|
|
|
|
# network
|
|
|
|
yq eval '.network // ""' ${HOSTFS}/etc/kubernetes/kubezero.yaml > _values.yaml
|
|
|
|
helm template kubezero/kubezero-network --version 0.1.0 --include-crds --namespace kube-system --name-template network \
|
|
|
|
-f _values.yaml --kube-version $KUBE_VERSION | kubectl apply -f - $LOG
|
|
|
|
|
|
|
|
# addons
|
|
|
|
yq eval '.addons // ""' ${HOSTFS}/etc/kubernetes/kubezero.yaml > _values.yaml
|
2021-12-03 21:13:40 +00:00
|
|
|
helm template kubezero/kubezero-addons --version 0.2.2 --include-crds --namespace kube-system --name-template addons \
|
2021-12-01 12:33:11 +00:00
|
|
|
-f _values.yaml --kube-version $KUBE_VERSION | kubectl apply -f - $LOG
|
|
|
|
fi
|
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
post_kubeadm
|
2021-12-01 12:33:11 +00:00
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
echo "${1} cluster $CLUSTERNAME successfull."
|
|
|
|
|
|
|
|
|
|
|
|
# Since 1.21 we only need to backup etcd + /etc/kubernetes/pki !
|
|
|
|
elif [ "$1" == 'backup' ]; then
|
|
|
|
mkdir -p ${WORKDIR}
|
|
|
|
|
|
|
|
restic snapshots || restic init || exit 1
|
|
|
|
|
2021-12-15 22:19:15 +00:00
|
|
|
etcdctl --endpoints=https://${ETCD_NODENAME}:2379 snapshot save ${WORKDIR}/etcd_snapshot
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
# pki & cluster-admin access
|
|
|
|
cp -r ${HOSTFS}/etc/kubernetes/pki ${WORKDIR}
|
|
|
|
cp -r ${HOSTFS}/etc/kubernetes/admin.conf ${WORKDIR}
|
2021-12-01 12:33:11 +00:00
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
# Backup via restic
|
|
|
|
restic snapshots || restic init
|
2021-12-03 21:13:40 +00:00
|
|
|
restic backup ${WORKDIR} -H $CLUSTERNAME --tag $VERSION
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
echo "Backup complete"
|
2021-12-03 21:13:40 +00:00
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
# Remove backups from previous versions
|
2021-12-03 21:13:40 +00:00
|
|
|
restic forget --keep-tag $VERSION --prune
|
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
# Regular retention
|
2021-11-27 13:02:23 +00:00
|
|
|
restic forget --keep-hourly 24 --keep-daily ${RESTIC_RETENTION:-7} --prune
|
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
# Defrag etcd backend
|
|
|
|
etcdctl --endpoints=https://${ETCD_NODENAME}:2379 defrag
|
|
|
|
|
2021-12-03 21:13:40 +00:00
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
elif [ "$1" == 'restore' ]; then
|
|
|
|
mkdir -p ${WORKDIR}
|
|
|
|
|
2021-12-03 21:13:40 +00:00
|
|
|
restic restore latest --no-lock -t / --tag $VERSION
|
2021-11-27 13:02:23 +00:00
|
|
|
|
|
|
|
# Make last etcd snapshot available
|
|
|
|
cp ${WORKDIR}/etcd_snapshot ${HOSTFS}/etc/kubernetes
|
|
|
|
|
|
|
|
# Put PKI in place
|
|
|
|
cp -r ${WORKDIR}/pki ${HOSTFS}/etc/kubernetes
|
|
|
|
|
|
|
|
# Always use kubeadm kubectl config to never run into chicken egg with custom auth hooks
|
|
|
|
cp ${WORKDIR}/admin.conf ${HOSTFS}/root/.kube/config
|
|
|
|
|
2021-12-17 10:57:23 +00:00
|
|
|
elif [ "$1" == 'debug_shell' ]; then
|
|
|
|
echo "Entering debug shell"
|
|
|
|
/bin/sh
|
|
|
|
|
2021-11-27 13:02:23 +00:00
|
|
|
else
|
|
|
|
echo "Unknown command!"
|
|
|
|
exit 1
|
|
|
|
fi
|