Merge pull request 'Automated Manifest Update' (#6271) from auto/update-manifests into manifests
Reviewed-on: #6271
This commit was merged in pull request #6271.
This commit is contained in:
@@ -55,7 +55,7 @@ spec:
|
|||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ntfy-url
|
key: ntfy-url
|
||||||
name: audiobookshelf-config-apprise
|
name: audiobookshelf-config-apprise
|
||||||
image: ghcr.io/caronc/apprise:v1.3.3@sha256:4bfeac268ba87b8e08e308c9aa0182fe99e9501ec464027afc333d1634e65977
|
image: ghcr.io/caronc/apprise:v1.4.0@sha256:9d97a6b9b42cf6afdf3b5466dbed2a59cd42a4bb777ec6aa57b5f2ee623569eb
|
||||||
name: apprise-api
|
name: apprise-api
|
||||||
- env:
|
- env:
|
||||||
- name: TZ
|
- name: TZ
|
||||||
|
|||||||
@@ -1,14 +1,15 @@
|
|||||||
apiVersion: batch/v1
|
apiVersion: batch/v1
|
||||||
kind: CronJob
|
kind: CronJob
|
||||||
metadata:
|
metadata:
|
||||||
name: rclone-directus-assets
|
name: directus-directus-assets-rclone
|
||||||
labels:
|
labels:
|
||||||
app.kubernetes.io/controller: directus-assets
|
app.kubernetes.io/controller: main
|
||||||
app.kubernetes.io/instance: rclone
|
app.kubernetes.io/instance: directus
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: rclone
|
app.kubernetes.io/name: directus-assets-rclone
|
||||||
helm.sh/chart: rclone-4.6.2
|
app.kubernetes.io/version: v1.73.5
|
||||||
namespace: rclone
|
helm.sh/chart: rclone-directus-assets-remote-0.2.0
|
||||||
|
namespace: directus
|
||||||
spec:
|
spec:
|
||||||
suspend: false
|
suspend: false
|
||||||
concurrencyPolicy: Forbid
|
concurrencyPolicy: Forbid
|
||||||
@@ -24,9 +25,9 @@ spec:
|
|||||||
template:
|
template:
|
||||||
metadata:
|
metadata:
|
||||||
labels:
|
labels:
|
||||||
app.kubernetes.io/controller: directus-assets
|
app.kubernetes.io/controller: main
|
||||||
app.kubernetes.io/instance: rclone
|
app.kubernetes.io/instance: directus
|
||||||
app.kubernetes.io/name: rclone
|
app.kubernetes.io/name: directus-assets-rclone
|
||||||
spec:
|
spec:
|
||||||
enableServiceLinks: false
|
enableServiceLinks: false
|
||||||
serviceAccountName: default
|
serviceAccountName: default
|
||||||
@@ -56,22 +57,22 @@ spec:
|
|||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ACCESS_KEY_ID
|
key: ACCESS_KEY_ID
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-source-config
|
||||||
- name: RCLONE_CONFIG_SRC_SECRET_ACCESS_KEY
|
- name: RCLONE_CONFIG_SRC_SECRET_ACCESS_KEY
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ACCESS_SECRET_KEY
|
key: ACCESS_SECRET_KEY
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-source-config
|
||||||
- name: RCLONE_CONFIG_SRC_REGION
|
- name: RCLONE_CONFIG_SRC_REGION
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ACCESS_REGION
|
key: ACCESS_REGION
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-source-config
|
||||||
- name: RCLONE_CONFIG_SRC_ENDPOINT
|
- name: RCLONE_CONFIG_SRC_ENDPOINT
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: SRC_ENDPOINT
|
key: SRC_ENDPOINT
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-source-config
|
||||||
- name: RCLONE_CONFIG_SRC_S3_FORCE_PATH_STYLE
|
- name: RCLONE_CONFIG_SRC_S3_FORCE_PATH_STYLE
|
||||||
value: "true"
|
value: "true"
|
||||||
- name: RCLONE_CONFIG_DEST_TYPE
|
- name: RCLONE_CONFIG_DEST_TYPE
|
||||||
@@ -84,23 +85,24 @@ spec:
|
|||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ACCESS_KEY_ID
|
key: ACCESS_KEY_ID
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-destination-config
|
||||||
- name: RCLONE_CONFIG_DEST_SECRET_ACCESS_KEY
|
- name: RCLONE_CONFIG_DEST_SECRET_ACCESS_KEY
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ACCESS_SECRET_KEY
|
key: ACCESS_SECRET_KEY
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-destination-config
|
||||||
- name: RCLONE_CONFIG_DEST_REGION
|
- name: RCLONE_CONFIG_DEST_REGION
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ACCESS_REGION
|
key: ACCESS_REGION
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-destination-config
|
||||||
- name: RCLONE_CONFIG_DEST_ENDPOINT
|
- name: RCLONE_CONFIG_DEST_ENDPOINT
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: DEST_ENDPOINT
|
key: DEST_ENDPOINT
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-destination-config
|
||||||
- name: RCLONE_CONFIG_DEST_S3_FORCE_PATH_STYLE
|
- name: RCLONE_CONFIG_SRC_DEST_FORCE_PATH_STYLE
|
||||||
value: "true"
|
value: "true"
|
||||||
image: rclone/rclone:1.73.5@sha256:1619a625f845e169c34b952cf40c483c0392965b821c5155cde8cbfd35254a96
|
image: rclone/rclone:1.73.5@sha256:1619a625f845e169c34b952cf40c483c0392965b821c5155cde8cbfd35254a96
|
||||||
|
imagePullPolicy: IfNotPresent
|
||||||
name: sync
|
name: sync
|
||||||
@@ -0,0 +1,33 @@
|
|||||||
|
apiVersion: external-secrets.io/v1
|
||||||
|
kind: ExternalSecret
|
||||||
|
metadata:
|
||||||
|
name: directus-assets-rclone-destination-config
|
||||||
|
namespace: directus
|
||||||
|
labels:
|
||||||
|
helm.sh/chart: rclone-directus-assets-remote-0.2.0
|
||||||
|
app.kubernetes.io/instance: directus
|
||||||
|
app.kubernetes.io/part-of: directus
|
||||||
|
app.kubernetes.io/version: "0.2.0"
|
||||||
|
app.kubernetes.io/managed-by: Helm
|
||||||
|
app.kubernetes.io/name: directus-assets-rclone-destination-config
|
||||||
|
spec:
|
||||||
|
secretStoreRef:
|
||||||
|
kind: ClusterSecretStore
|
||||||
|
name: openbao
|
||||||
|
data:
|
||||||
|
- secretKey: ACCESS_KEY_ID
|
||||||
|
remoteRef:
|
||||||
|
key: /garage/home-infra/directus-assets
|
||||||
|
property: ACCESS_KEY_ID
|
||||||
|
- secretKey: ACCESS_SECRET_KEY
|
||||||
|
remoteRef:
|
||||||
|
key: /garage/home-infra/directus-assets
|
||||||
|
property: ACCESS_KEY_ID
|
||||||
|
- secretKey: ACCESS_REGION
|
||||||
|
remoteRef:
|
||||||
|
key: /garage/home-infra/directus-assets
|
||||||
|
property: ACCESS_KEY_ID
|
||||||
|
- secretKey: DEST_ENDPOINT
|
||||||
|
remoteRef:
|
||||||
|
key: /garage/config
|
||||||
|
property: ENDPOINT_REMOTE
|
||||||
@@ -1,12 +1,15 @@
|
|||||||
apiVersion: external-secrets.io/v1
|
apiVersion: external-secrets.io/v1
|
||||||
kind: ExternalSecret
|
kind: ExternalSecret
|
||||||
metadata:
|
metadata:
|
||||||
name: garage-directus-secret
|
name: directus-assets-rclone-source-config
|
||||||
namespace: rclone
|
namespace: directus
|
||||||
labels:
|
labels:
|
||||||
app.kubernetes.io/name: garage-directus-secret
|
helm.sh/chart: rclone-directus-assets-remote-0.2.0
|
||||||
app.kubernetes.io/instance: rclone
|
app.kubernetes.io/instance: directus
|
||||||
app.kubernetes.io/part-of: rclone
|
app.kubernetes.io/part-of: directus
|
||||||
|
app.kubernetes.io/version: "0.2.0"
|
||||||
|
app.kubernetes.io/managed-by: Helm
|
||||||
|
app.kubernetes.io/name: directus-assets-rclone-source-config
|
||||||
spec:
|
spec:
|
||||||
secretStoreRef:
|
secretStoreRef:
|
||||||
kind: ClusterSecretStore
|
kind: ClusterSecretStore
|
||||||
@@ -16,19 +19,15 @@ spec:
|
|||||||
remoteRef:
|
remoteRef:
|
||||||
key: /garage/home-infra/directus-assets
|
key: /garage/home-infra/directus-assets
|
||||||
property: ACCESS_KEY_ID
|
property: ACCESS_KEY_ID
|
||||||
- secretKey: ACCESS_REGION
|
|
||||||
remoteRef:
|
|
||||||
key: /garage/home-infra/directus-assets
|
|
||||||
property: ACCESS_REGION
|
|
||||||
- secretKey: ACCESS_SECRET_KEY
|
- secretKey: ACCESS_SECRET_KEY
|
||||||
remoteRef:
|
remoteRef:
|
||||||
key: /garage/home-infra/directus-assets
|
key: /garage/home-infra/directus-assets
|
||||||
property: ACCESS_SECRET_KEY
|
property: ACCESS_SECRET_KEY
|
||||||
|
- secretKey: ACCESS_REGION
|
||||||
|
remoteRef:
|
||||||
|
key: /garage/home-infra/directus-assets
|
||||||
|
property: ACCESS_REGION
|
||||||
- secretKey: SRC_ENDPOINT
|
- secretKey: SRC_ENDPOINT
|
||||||
remoteRef:
|
remoteRef:
|
||||||
key: /garage/config
|
key: /garage/config
|
||||||
property: ENDPOINT_LOCAL
|
property: ENDPOINT_LOCAL
|
||||||
- secretKey: DEST_ENDPOINT
|
|
||||||
remoteRef:
|
|
||||||
key: /garage/config
|
|
||||||
property: ENDPOINT_REMOTE
|
|
||||||
@@ -50,24 +50,24 @@ spec:
|
|||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: navidrome-user
|
key: navidrome-user
|
||||||
name: music-grabber-config-secret
|
name: music-grabber-config
|
||||||
- name: NAVIDROME_PASS
|
- name: NAVIDROME_PASS
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: navidrome-password
|
key: navidrome-password
|
||||||
name: music-grabber-config-secret
|
name: music-grabber-config
|
||||||
- name: SLSKD_URL
|
- name: SLSKD_URL
|
||||||
value: http://slskd.slskd:5030
|
value: http://slskd.slskd:5030
|
||||||
- name: SLSKD_USER
|
- name: SLSKD_USER
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: slskd-user
|
key: slskd-user
|
||||||
name: music-grabber-config-secret
|
name: music-grabber-config
|
||||||
- name: SLSKD_PASS
|
- name: SLSKD_PASS
|
||||||
valueFrom:
|
valueFrom:
|
||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: slskd-password
|
key: slskd-password
|
||||||
name: music-grabber-config-secret
|
name: music-grabber-config
|
||||||
- name: SLSKD_DOWNLOADS_PATH
|
- name: SLSKD_DOWNLOADS_PATH
|
||||||
value: /mnt/store/slskd/Downloads
|
value: /mnt/store/slskd/Downloads
|
||||||
image: g33kphr33k/musicgrabber:2.6.6@sha256:dad8dec4e32671ef7326d31f58ea626fa4622571e65c6bb34459bc2648f1fead
|
image: g33kphr33k/musicgrabber:2.6.6@sha256:dad8dec4e32671ef7326d31f58ea626fa4622571e65c6bb34459bc2648f1fead
|
||||||
|
|||||||
@@ -72,7 +72,7 @@ spec:
|
|||||||
secretKeyRef:
|
secretKeyRef:
|
||||||
key: ntfy-url
|
key: ntfy-url
|
||||||
name: qbit-manage-config
|
name: qbit-manage-config
|
||||||
image: ghcr.io/caronc/apprise:v1.3.3@sha256:4bfeac268ba87b8e08e308c9aa0182fe99e9501ec464027afc333d1634e65977
|
image: ghcr.io/caronc/apprise:v1.4.0@sha256:9d97a6b9b42cf6afdf3b5466dbed2a59cd42a4bb777ec6aa57b5f2ee623569eb
|
||||||
name: apprise-api
|
name: apprise-api
|
||||||
- env:
|
- env:
|
||||||
- name: TZ
|
- name: TZ
|
||||||
|
|||||||
@@ -29,7 +29,7 @@ spec:
|
|||||||
app.kubernetes.io/name: talos
|
app.kubernetes.io/name: talos
|
||||||
spec:
|
spec:
|
||||||
enableServiceLinks: false
|
enableServiceLinks: false
|
||||||
serviceAccountName: default
|
serviceAccountName: talos-defrag
|
||||||
automountServiceAccountToken: true
|
automountServiceAccountToken: true
|
||||||
hostIPC: false
|
hostIPC: false
|
||||||
hostNetwork: false
|
hostNetwork: false
|
||||||
|
|||||||
@@ -29,7 +29,7 @@ spec:
|
|||||||
app.kubernetes.io/name: talos
|
app.kubernetes.io/name: talos
|
||||||
spec:
|
spec:
|
||||||
enableServiceLinks: false
|
enableServiceLinks: false
|
||||||
serviceAccountName: default
|
serviceAccountName: talos-defrag
|
||||||
automountServiceAccountToken: true
|
automountServiceAccountToken: true
|
||||||
hostIPC: false
|
hostIPC: false
|
||||||
hostNetwork: false
|
hostNetwork: false
|
||||||
|
|||||||
@@ -29,7 +29,7 @@ spec:
|
|||||||
app.kubernetes.io/name: talos
|
app.kubernetes.io/name: talos
|
||||||
spec:
|
spec:
|
||||||
enableServiceLinks: false
|
enableServiceLinks: false
|
||||||
serviceAccountName: default
|
serviceAccountName: talos-defrag
|
||||||
automountServiceAccountToken: true
|
automountServiceAccountToken: true
|
||||||
hostIPC: false
|
hostIPC: false
|
||||||
hostNetwork: false
|
hostNetwork: false
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ spec:
|
|||||||
provider: openbao
|
provider: openbao
|
||||||
parameters:
|
parameters:
|
||||||
baoAddress: "http://openbao-internal.openbao:8200"
|
baoAddress: "http://openbao-internal.openbao:8200"
|
||||||
roleName: talos-backup
|
roleName: talos-defrag
|
||||||
objects: |
|
objects: |
|
||||||
- objectName: config
|
- objectName: config
|
||||||
fileName: config
|
fileName: config
|
||||||
|
|||||||
@@ -0,0 +1,10 @@
|
|||||||
|
apiVersion: v1
|
||||||
|
kind: ServiceAccount
|
||||||
|
metadata:
|
||||||
|
name: etcd-defrag
|
||||||
|
labels:
|
||||||
|
app.kubernetes.io/instance: talos
|
||||||
|
app.kubernetes.io/managed-by: Helm
|
||||||
|
app.kubernetes.io/name: talos
|
||||||
|
helm.sh/chart: etcd-defrag-4.6.2
|
||||||
|
namespace: talos
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
apiVersion: v1
|
|
||||||
kind: ServiceAccount
|
|
||||||
metadata:
|
|
||||||
name: talos-backup
|
|
||||||
namespace: talos
|
|
||||||
labels:
|
|
||||||
app.kubernetes.io/name: talos-backup
|
|
||||||
app.kubernetes.io/instance: talos
|
|
||||||
app.kubernetes.io/part-of: talos
|
|
||||||
10
clusters/cl01tl/manifests/talos/ServiceAccount-talos.yaml
Normal file
10
clusters/cl01tl/manifests/talos/ServiceAccount-talos.yaml
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
apiVersion: v1
|
||||||
|
kind: ServiceAccount
|
||||||
|
metadata:
|
||||||
|
name: talos
|
||||||
|
labels:
|
||||||
|
app.kubernetes.io/instance: talos
|
||||||
|
app.kubernetes.io/managed-by: Helm
|
||||||
|
app.kubernetes.io/name: talos
|
||||||
|
helm.sh/chart: etcd-backup-4.6.2
|
||||||
|
namespace: talos
|
||||||
@@ -12,75 +12,3 @@ data:
|
|||||||
echo " ";
|
echo " ";
|
||||||
echo ">> Running S3 backup for Vault snapshot";
|
echo ">> Running S3 backup for Vault snapshot";
|
||||||
OUTPUT=$(s3cmd sync --no-check-certificate -v /opt/backup/* "${BUCKET}/cl01tl/cl01tl-vault-snapshots/" 2>&1)
|
OUTPUT=$(s3cmd sync --no-check-certificate -v /opt/backup/* "${BUCKET}/cl01tl/cl01tl-vault-snapshots/" 2>&1)
|
||||||
STATUS=$?
|
|
||||||
|
|
||||||
if [ $STATUS -ne 0 ]; then
|
|
||||||
if echo "$OUTPUT" | grep -q "403 Forbidden"; then
|
|
||||||
MESSAGE="403 Authentication Error: Your keys are wrong or you don't have permission"
|
|
||||||
elif echo "$OUTPUT" | grep -q "404 Not Found"; then
|
|
||||||
MESSAGE="404 Error: The bucket or folder does not exist"
|
|
||||||
elif echo "$OUTPUT" | grep -q "Connection refused"; then
|
|
||||||
MESSAGE="Network Error: Cannot reach the S3 endpoint"
|
|
||||||
else
|
|
||||||
MESSAGE="Unknown Error"
|
|
||||||
echo " ";
|
|
||||||
echo ">> Unknown Error, output:"
|
|
||||||
echo " "
|
|
||||||
echo "$OUTPUT"
|
|
||||||
fi
|
|
||||||
|
|
||||||
MAX_RETRIES=5
|
|
||||||
SUCCESS=false
|
|
||||||
|
|
||||||
echo " "
|
|
||||||
echo ">> Sending message to ntfy using curl ..."
|
|
||||||
|
|
||||||
echo " "
|
|
||||||
echo ">> Verifying required commands ..."
|
|
||||||
|
|
||||||
for i in $(seq 1 "$MAX_RETRIES"); do
|
|
||||||
if apk update 2>&1 >/dev/null; then
|
|
||||||
echo ">> Attempt $i: Repositories are reachable";
|
|
||||||
SUCCESS=true;
|
|
||||||
break;
|
|
||||||
else
|
|
||||||
echo ">> Attempt $i: Connection failed, retrying in 5 seconds ...";
|
|
||||||
sleep 5;
|
|
||||||
fi;
|
|
||||||
done;
|
|
||||||
|
|
||||||
if [ "$SUCCESS" = false ]; then
|
|
||||||
echo ">> ERROR: Could not connect to apk repositories after $MAX_RETRIES attempts, exiting ...";
|
|
||||||
exit 1;
|
|
||||||
fi
|
|
||||||
|
|
||||||
if ! command -v curl 2>&1 >/dev/null; then
|
|
||||||
echo ">> Command curl could not be found, installing";
|
|
||||||
apk add --no-cache -q curl;
|
|
||||||
if [ $? -eq 0 ]; then
|
|
||||||
echo ">> Installation successful";
|
|
||||||
else
|
|
||||||
echo ">> Installation failed with exit code $?";
|
|
||||||
exit 1;
|
|
||||||
fi;
|
|
||||||
fi;
|
|
||||||
|
|
||||||
echo " "
|
|
||||||
echo ">> Sending to NTFY ..."
|
|
||||||
echo ">> Message: $MESSAGE"
|
|
||||||
HTTP_STATUS=$(curl \
|
|
||||||
--silent \
|
|
||||||
--write-out '%{http_code}' \
|
|
||||||
-H "Authorization: Bearer ${NTFY_TOKEN}" \
|
|
||||||
-H "X-Priority: 5" \
|
|
||||||
-H "X-Tags: warning" \
|
|
||||||
-H "X-Title: Vault Backup Failed for ${TARGET}" \
|
|
||||||
-d "$MESSAGE" \
|
|
||||||
${NTFY_ENDPOINT}/${NTFY_TOPIC}
|
|
||||||
)
|
|
||||||
echo ">> HTTP Status Code: $HTTP_STATUS"
|
|
||||||
|
|
||||||
else
|
|
||||||
echo " ";
|
|
||||||
echo ">> S3 Sync succeeded"
|
|
||||||
fi
|
|
||||||
|
|||||||
@@ -10,56 +10,26 @@ metadata:
|
|||||||
data:
|
data:
|
||||||
snapshot.sh: |
|
snapshot.sh: |
|
||||||
DATE=$(date +"%Y%m%d-%H-%M")
|
DATE=$(date +"%Y%m%d-%H-%M")
|
||||||
MAX_RETRIES=5
|
|
||||||
SUCCESS=false
|
|
||||||
|
|
||||||
echo " "
|
echo " "
|
||||||
echo ">> Running Vault Snapshot Script ..."
|
echo ">> Running Vault Snapshot Script ..."
|
||||||
|
|
||||||
echo " "
|
echo " "
|
||||||
echo ">> Verifying required commands ..."
|
echo ">> Fetching Vault token ..."
|
||||||
echo " "
|
export VAULT_TOKEN=$(vault write -field=token auth/approle/login role_id=$VAULT_APPROLE_ROLE_ID secret_id=$VAULT_APPROLE_SECRET_ID)
|
||||||
|
|
||||||
for i in $(seq 1 "$MAX_RETRIES"); do
|
if [ -z "$VAULT_TOKEN" ]; then
|
||||||
if apk update 2>&1 >/dev/null; then
|
echo ">> ERROR: Failed to fetch Vault token! Exiting..."
|
||||||
echo ">> Attempt $i: Repositories are reachable";
|
exit 1
|
||||||
SUCCESS=true;
|
|
||||||
break;
|
|
||||||
else
|
|
||||||
echo ">> Attempt $i: Connection failed, retrying in 5 seconds ...";
|
|
||||||
sleep 5;
|
|
||||||
fi;
|
|
||||||
done;
|
|
||||||
|
|
||||||
if [ "$SUCCESS" = false ]; then
|
|
||||||
echo ">> ERROR: Could not connect to apk repositories after $MAX_RETRIES attempts, exiting ...";
|
|
||||||
exit 1;
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
echo " "
|
echo " "
|
||||||
|
echo ">> Taking Vault snapshot ..."
|
||||||
if ! command -v jq 2>&1 >/dev/null; then
|
|
||||||
echo ">> Command jq could not be found, installing";
|
|
||||||
apk add --no-cache -q jq;
|
|
||||||
if [ $? -eq 0 ]; then
|
|
||||||
echo ">> Installation successful";
|
|
||||||
else
|
|
||||||
echo ">> Installation failed with exit code $?";
|
|
||||||
exit 1;
|
|
||||||
fi;
|
|
||||||
fi;
|
|
||||||
|
|
||||||
echo " ";
|
|
||||||
echo ">> Fetching Vault token ...";
|
|
||||||
export VAULT_TOKEN=$(vault write auth/approle/login role_id=$VAULT_APPROLE_ROLE_ID secret_id=$VAULT_APPROLE_SECRET_ID -format=json | jq -r .auth.client_token);
|
|
||||||
|
|
||||||
echo " ";
|
|
||||||
echo ">> Taking Vault snapsot ...";
|
|
||||||
vault operator raft snapshot save /opt/backup/vault-snapshot-$DATE.snap
|
vault operator raft snapshot save /opt/backup/vault-snapshot-$DATE.snap
|
||||||
|
|
||||||
echo " ";
|
echo " "
|
||||||
echo ">> Setting ownership of Vault snapsot ...";
|
echo ">> Setting ownership of Vault snapshot ..."
|
||||||
chown 100:1000 /opt/backup/vault-snapshot-$DATE.snap
|
chown 100:1000 /opt/backup/vault-snapshot-$DATE.snap
|
||||||
|
|
||||||
echo " ";
|
echo " "
|
||||||
echo ">> Completed Vault snapshot";
|
echo ">> Completed Vault snapshot"
|
||||||
|
|||||||
Reference in New Issue
Block a user