Sample logical backup image. (#572)
* Sample logical backup image. Based on the earlier work by Dmitry Dolgov @erthalion
This commit is contained in:
parent
b619569e28
commit
3ffc8ac5fa
|
|
@ -0,0 +1,33 @@
|
||||||
|
FROM ubuntu:18.04
|
||||||
|
LABEL maintainer="Team ACID @ Zalando <team-acid@zalando.de>"
|
||||||
|
|
||||||
|
SHELL ["/bin/bash", "-o", "pipefail", "-c"]
|
||||||
|
RUN apt-get update \
|
||||||
|
&& apt-get install --no-install-recommends -y \
|
||||||
|
apt-utils \
|
||||||
|
ca-certificates \
|
||||||
|
lsb-release \
|
||||||
|
pigz \
|
||||||
|
python3-pip \
|
||||||
|
python3-setuptools \
|
||||||
|
curl \
|
||||||
|
jq \
|
||||||
|
gnupg \
|
||||||
|
&& pip3 install --no-cache-dir awscli --upgrade \
|
||||||
|
&& echo "deb http://apt.postgresql.org/pub/repos/apt/ $(lsb_release -cs)-pgdg main" > /etc/apt/sources.list.d/pgdg.list \
|
||||||
|
&& cat /etc/apt/sources.list.d/pgdg.list \
|
||||||
|
&& curl --silent https://www.postgresql.org/media/keys/ACCC4CF8.asc | apt-key add - \
|
||||||
|
&& apt-get update \
|
||||||
|
&& apt-get install --no-install-recommends -y \
|
||||||
|
postgresql-client-11 \
|
||||||
|
postgresql-client-10 \
|
||||||
|
postgresql-client-9.6 \
|
||||||
|
postgresql-client-9.5 \
|
||||||
|
&& apt-get clean \
|
||||||
|
&& rm -rf /var/lib/apt/lists/*
|
||||||
|
|
||||||
|
COPY dump.sh ./
|
||||||
|
|
||||||
|
ENV PG_DIR=/usr/lib/postgresql/
|
||||||
|
|
||||||
|
ENTRYPOINT ["/dump.sh"]
|
||||||
|
|
@ -0,0 +1,94 @@
|
||||||
|
#! /usr/bin/env bash
|
||||||
|
|
||||||
|
# enable unofficial bash strict mode
|
||||||
|
set -o errexit
|
||||||
|
set -o nounset
|
||||||
|
set -o pipefail
|
||||||
|
IFS=$'\n\t'
|
||||||
|
|
||||||
|
# make script trace visible via `kubectl logs`
|
||||||
|
set -o xtrace
|
||||||
|
|
||||||
|
ALL_DB_SIZE_QUERY="select sum(pg_database_size(datname)::numeric) from pg_database;"
|
||||||
|
PG_BIN=$PG_DIR/$PG_VERSION/bin
|
||||||
|
DUMP_SIZE_COEFF=5
|
||||||
|
|
||||||
|
TOKEN=$(cat /var/run/secrets/kubernetes.io/serviceaccount/token)
|
||||||
|
K8S_API_URL=https://$KUBERNETES_SERVICE_HOST:$KUBERNETES_SERVICE_PORT/api/v1
|
||||||
|
CERT=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt
|
||||||
|
|
||||||
|
function estimate_size {
|
||||||
|
"$PG_BIN"/psql -tqAc "${ALL_DB_SIZE_QUERY}"
|
||||||
|
}
|
||||||
|
|
||||||
|
function dump {
|
||||||
|
# settings are taken from the environment
|
||||||
|
"$PG_BIN"/pg_dumpall
|
||||||
|
}
|
||||||
|
|
||||||
|
function compress {
|
||||||
|
pigz
|
||||||
|
}
|
||||||
|
|
||||||
|
function aws_upload {
|
||||||
|
declare -r EXPECTED_SIZE="$1"
|
||||||
|
|
||||||
|
# mimic bucket setup from Spilo
|
||||||
|
# to keep logical backups at the same path as WAL
|
||||||
|
# NB: $LOGICAL_BACKUP_S3_BUCKET_SCOPE_SUFFIX already contains the leading "/" when set by the Postgres operator
|
||||||
|
PATH_TO_BACKUP=s3://$LOGICAL_BACKUP_S3_BUCKET"/spilo/"$SCOPE$LOGICAL_BACKUP_S3_BUCKET_SCOPE_SUFFIX"/logical_backups/"$(date +%s).sql.gz
|
||||||
|
|
||||||
|
if [ -z "$EXPECTED_SIZE" ]; then
|
||||||
|
aws s3 cp - "$PATH_TO_BACKUP" --debug --sse="AES256"
|
||||||
|
else
|
||||||
|
aws s3 cp - "$PATH_TO_BACKUP" --debug --expected-size "$EXPECTED_SIZE" --sse="AES256"
|
||||||
|
fi;
|
||||||
|
}
|
||||||
|
|
||||||
|
function get_pods {
|
||||||
|
declare -r SELECTOR="$1"
|
||||||
|
|
||||||
|
curl "${K8S_API_URL}/pods?$SELECTOR" \
|
||||||
|
--cacert $CERT \
|
||||||
|
-H "Authorization: Bearer ${TOKEN}" | jq .items[].status.podIP -r
|
||||||
|
}
|
||||||
|
|
||||||
|
function get_current_pod {
|
||||||
|
curl "${K8S_API_URL}/pods?fieldSelector=metadata.name%3D${HOSTNAME}" \
|
||||||
|
--cacert $CERT \
|
||||||
|
-H "Authorization: Bearer ${TOKEN}"
|
||||||
|
}
|
||||||
|
|
||||||
|
declare -a search_strategy=(
|
||||||
|
list_all_replica_pods_current_node
|
||||||
|
list_all_replica_pods_any_node
|
||||||
|
get_master_pod
|
||||||
|
)
|
||||||
|
|
||||||
|
function list_all_replica_pods_current_node {
|
||||||
|
get_pods "labelSelector=version%3D${SCOPE},spilo-role%3Dreplica&fieldSelector=spec.nodeName%3D${CURRENT_NODENAME}" | head -n 1
|
||||||
|
}
|
||||||
|
|
||||||
|
function list_all_replica_pods_any_node {
|
||||||
|
get_pods "labelSelector=version%3D${SCOPE},spilo-role%3Dreplica" | head -n 1
|
||||||
|
}
|
||||||
|
|
||||||
|
function get_master_pod {
|
||||||
|
get_pods "labelSelector=version%3D${SCOPE},spilo-role%3Dmaster" | head -n 1
|
||||||
|
}
|
||||||
|
|
||||||
|
CURRENT_NODENAME=$(get_current_pod | jq .items[].spec.nodeName --raw-output)
|
||||||
|
export CURRENT_NODENAME
|
||||||
|
|
||||||
|
for search in "${search_strategy[@]}"; do
|
||||||
|
|
||||||
|
PGHOST=$(eval "$search")
|
||||||
|
export PGHOST
|
||||||
|
|
||||||
|
if [ -n "$PGHOST" ]; then
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
|
||||||
|
done
|
||||||
|
|
||||||
|
dump | compress | aws_upload $(($(estimate_size) / DUMP_SIZE_COEFF))
|
||||||
|
|
@ -346,7 +346,7 @@ The operator logs reasons for a rolling update with the `info` level and a diff
|
||||||
|
|
||||||
The operator can manage k8s cron jobs to run logical backups of Postgres clusters. The cron job periodically spawns a batch job that runs a single pod. The backup script within this pod's container can connect to a DB for a logical backup. The operator updates cron jobs during Sync if the job schedule changes; the job name acts as the job identifier. These jobs are to be enabled for each indvidual Postgres cluster by setting `enableLogicalBackup: true` in its manifest. Notes:
|
The operator can manage k8s cron jobs to run logical backups of Postgres clusters. The cron job periodically spawns a batch job that runs a single pod. The backup script within this pod's container can connect to a DB for a logical backup. The operator updates cron jobs during Sync if the job schedule changes; the job name acts as the job identifier. These jobs are to be enabled for each indvidual Postgres cluster by setting `enableLogicalBackup: true` in its manifest. Notes:
|
||||||
|
|
||||||
1. The provided `registry.opensource.zalan.do/acid/logical-backup` image implements the backup via `pg_dumpall` and upload of (compressed) results to an S3 bucket; `pg_dumpall` requires a `superuser` access to a DB and runs on the replica when possible.
|
1. The [example image](../docker/logical-backup/Dockerfile) implements the backup via `pg_dumpall` and upload of compressed and encrypted results to an S3 bucket; the default image ``registry.opensource.zalan.do/acid/logical-backup`` is the same image built with the Zalando-internal CI pipeline. `pg_dumpall` requires a `superuser` access to a DB and runs on the replica when possible.
|
||||||
|
|
||||||
2. Due to the [limitation of Kubernetes cron jobs](https://kubernetes.io/docs/concepts/workloads/controllers/cron-jobs/#cron-job-limitations) it is highly advisable to set up additional monitoring for this feature; such monitoring is outside of the scope of operator responsibilities.
|
2. Due to the [limitation of Kubernetes cron jobs](https://kubernetes.io/docs/concepts/workloads/controllers/cron-jobs/#cron-job-limitations) it is highly advisable to set up additional monitoring for this feature; such monitoring is outside of the scope of operator responsibilities.
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -485,9 +485,9 @@ scalyr sidecar. In the CRD-based configuration they are grouped under the
|
||||||
|
|
||||||
* **logical_backup_schedule**
|
* **logical_backup_schedule**
|
||||||
Backup schedule in the cron format. Please take [the reference schedule format](https://kubernetes.io/docs/tasks/job/automated-tasks-with-cron-jobs/#schedule) into account. Default: "30 00 \* \* \*"
|
Backup schedule in the cron format. Please take [the reference schedule format](https://kubernetes.io/docs/tasks/job/automated-tasks-with-cron-jobs/#schedule) into account. Default: "30 00 \* \* \*"
|
||||||
|
|
||||||
* **logical_backup_docker_image**
|
* **logical_backup_docker_image**
|
||||||
Docker image for the pods of the cron job. Must implement backup logic and correctly handle pod and job restarts. The default image runs `pg_dumpall` (on a replica if possible) and uploads compressed results to an S3 bucket under the key `/spilo/pg_cluster_name/cluster_k8s_uuid/logical_backups` Default: "registry.opensource.zalan.do/acid/logical-backup"
|
An image for pods of the logical backup job. The [example image](../../docker/logical-backup/Dockerfile) runs `pg_dumpall` on a replica if possible and uploads compressed results to an S3 bucket under the key `/spilo/pg_cluster_name/cluster_k8s_uuid/logical_backups`. The default image is the same image built with the Zalando-internal CI pipeline. Default: "registry.opensource.zalan.do/acid/logical-backup"
|
||||||
|
|
||||||
* **logical_backup_s3_bucket**
|
* **logical_backup_s3_bucket**
|
||||||
S3 bucket to store backup results. The bucket has to be present and accessible by Postgres pods. Default: empty.
|
S3 bucket to store backup results. The bucket has to be present and accessible by Postgres pods. Default: empty.
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue