Logging tunings #24
@ -2,7 +2,7 @@ apiVersion: v2
|
|||||||
name: kubezero-logging
|
name: kubezero-logging
|
||||||
description: KubeZero Umbrella Chart for complete EFK stack
|
description: KubeZero Umbrella Chart for complete EFK stack
|
||||||
type: application
|
type: application
|
||||||
version: 0.3.6
|
version: 0.3.8
|
||||||
appVersion: 1.2.1
|
appVersion: 1.2.1
|
||||||
home: https://kubezero.com
|
home: https://kubezero.com
|
||||||
icon: https://cdn.zero-downtime.net/assets/kubezero/logo-small-64.png
|
icon: https://cdn.zero-downtime.net/assets/kubezero/logo-small-64.png
|
||||||
|
@ -108,6 +108,8 @@ fluentd:
|
|||||||
port 24224
|
port 24224
|
||||||
bind 0.0.0.0
|
bind 0.0.0.0
|
||||||
skip_invalid_event true
|
skip_invalid_event true
|
||||||
|
# Only for TCP not TLS
|
||||||
|
# send_keepalive_packet true
|
||||||
<transport tls>
|
<transport tls>
|
||||||
cert_path /mnt/fluentd-certs/tls.crt
|
cert_path /mnt/fluentd-certs/tls.crt
|
||||||
private_key_path /mnt/fluentd-certs/tls.key
|
private_key_path /mnt/fluentd-certs/tls.key
|
||||||
@ -128,7 +130,6 @@ fluentd:
|
|||||||
remove_keys id
|
remove_keys id
|
||||||
|
|
||||||
# KubeZero pipeline incl. GeoIP etc.
|
# KubeZero pipeline incl. GeoIP etc.
|
||||||
# Freaking ES jams under load and all is lost ...
|
|
||||||
# pipeline fluentd
|
# pipeline fluentd
|
||||||
|
|
||||||
host "#{ENV['OUTPUT_HOST']}"
|
host "#{ENV['OUTPUT_HOST']}"
|
||||||
@ -146,8 +147,8 @@ fluentd:
|
|||||||
request_timeout 15s
|
request_timeout 15s
|
||||||
suppress_type_name true
|
suppress_type_name true
|
||||||
|
|
||||||
<buffer>
|
<buffer tag>
|
||||||
@type file
|
@type file_single
|
||||||
path /var/log/fluentd-buffers/kubernetes.system.buffer
|
path /var/log/fluentd-buffers/kubernetes.system.buffer
|
||||||
flush_mode interval
|
flush_mode interval
|
||||||
flush_thread_count 2
|
flush_thread_count 2
|
||||||
@ -155,7 +156,6 @@ fluentd:
|
|||||||
flush_at_shutdown true
|
flush_at_shutdown true
|
||||||
retry_type exponential_backoff
|
retry_type exponential_backoff
|
||||||
retry_timeout 60m
|
retry_timeout 60m
|
||||||
chunk_limit_size 16M
|
|
||||||
overflow_action drop_oldest_chunk
|
overflow_action drop_oldest_chunk
|
||||||
</buffer>
|
</buffer>
|
||||||
</match>
|
</match>
|
||||||
|
@ -2,6 +2,7 @@
|
|||||||
set -e
|
set -e
|
||||||
|
|
||||||
DEPLOY_DIR=$( dirname $( realpath $0 ))
|
DEPLOY_DIR=$( dirname $( realpath $0 ))
|
||||||
|
which yq || { echo "yq not found!"; exit 1; }
|
||||||
|
|
||||||
# Waits for max 300s and retries
|
# Waits for max 300s and retries
|
||||||
function wait_for() {
|
function wait_for() {
|
||||||
@ -31,13 +32,15 @@ else
|
|||||||
_argo_date="$(date -u --iso-8601=seconds)"
|
_argo_date="$(date -u --iso-8601=seconds)"
|
||||||
_argo_passwd="$($DEPLOY_DIR/argocd_password.py)"
|
_argo_passwd="$($DEPLOY_DIR/argocd_password.py)"
|
||||||
|
|
||||||
cat <<EOF >> values.yaml
|
cat <<EOF > _argocd_values.yaml
|
||||||
|
argo-cd:
|
||||||
configs:
|
configs:
|
||||||
secret:
|
secret:
|
||||||
# ArgoCD password: ${_argo_passwd%%:*} Please move to secure location !
|
# ArgoCD password: ${_argo_passwd%%:*} Please move to secure location !
|
||||||
argocdServerAdminPassword: "${_argo_passwd##*:}"
|
argocdServerAdminPassword: "${_argo_passwd##*:}"
|
||||||
argocdServerAdminPasswordMtime: "$_argo_date"
|
argocdServerAdminPasswordMtime: "$_argo_date"
|
||||||
EOF
|
EOF
|
||||||
|
yq merge -i --overwrite values.yaml _argocd_values.yaml && rm -f _argocd_values.yaml
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Deploy initial argocd
|
# Deploy initial argocd
|
||||||
@ -58,7 +61,6 @@ EOF
|
|||||||
helm template $DEPLOY_DIR -f values.yaml -f cloudbender.yaml -f $DEPLOY_DIR/values-step-2.yaml > generated-values.yaml
|
helm template $DEPLOY_DIR -f values.yaml -f cloudbender.yaml -f $DEPLOY_DIR/values-step-2.yaml > generated-values.yaml
|
||||||
helm upgrade -n argocd kubezero kubezero/kubezero-argo-cd -f generated-values.yaml
|
helm upgrade -n argocd kubezero kubezero/kubezero-argo-cd -f generated-values.yaml
|
||||||
wait_for kubectl get Issuer -n kube-system kubezero-local-ca-issuer 2>/dev/null 1>&2
|
wait_for kubectl get Issuer -n kube-system kubezero-local-ca-issuer 2>/dev/null 1>&2
|
||||||
wait_for kubectl get ClusterIssuer letsencrypt-dns-prod 2>/dev/null 1>&2
|
|
||||||
kubectl wait --for=condition=Ready -n kube-system Issuer/kubezero-local-ca-issuer
|
kubectl wait --for=condition=Ready -n kube-system Issuer/kubezero-local-ca-issuer
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user