Adding the support for standby cluster
This will set up a continuous wal streaming cluster, by adding the corresponding section in postgres manifest. Instead of having a full-fledged standby cluster as in Patroni, here we use only the wal path of the source cluster and stream from there. Since, standby cluster is streaming from the master and does not require to create or use databases of it's own. Hence, it bypasses the creation of users or databases. There is a separate sample manifest added to set up a standby-cluster.
This commit is contained in:
parent
93bfed3e75
commit
540d58d5bd
|
|
@ -199,6 +199,9 @@ explanation of `ttl` and `loop_wait` parameters.
|
||||||
automatically created by Patroni for cluster members and permanent replication
|
automatically created by Patroni for cluster members and permanent replication
|
||||||
slots. Optional.
|
slots. Optional.
|
||||||
|
|
||||||
|
* **standby**
|
||||||
|
initializes cluster as a standby creating a cascading replication, where standby leader is streaming from specified remote location
|
||||||
|
|
||||||
## Postgres container resources
|
## Postgres container resources
|
||||||
|
|
||||||
Those parameters define [CPU and memory requests and
|
Those parameters define [CPU and memory requests and
|
||||||
|
|
|
||||||
17
docs/user.md
17
docs/user.md
|
|
@ -281,6 +281,23 @@ spec:
|
||||||
s3_force_path_style: true
|
s3_force_path_style: true
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Setting up a standby cluster
|
||||||
|
|
||||||
|
Standby clusters are like normal cluster but they are streaming from a remote cluster. As the first version of this feature, the only scenario covered by operator is to stream from a wal archive of the master. Following the more popular infrastructure of using Amazon's S3 buckets, it is mentioned as s3_wal_path here. To make a cluster as standby add a section standby in the YAML file as follows.
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
spec:
|
||||||
|
standby:
|
||||||
|
s3_wal_path: "s3 bucket path to the master"
|
||||||
|
```
|
||||||
|
|
||||||
|
Things to note:
|
||||||
|
|
||||||
|
- An empty string is provided in s3_wal_path of the standby cluster will result in error and no statefulset will be created.
|
||||||
|
- Only one pod can be deployed for stand-by cluster.
|
||||||
|
- To manually promote the standby_cluster, use patronictl and remove config entry.
|
||||||
|
- There is no way to transform a non-standby cluster to standby cluster through operator. Hence, if a cluster is created without standby section in YAML and later modified by adding that section, there will be no effect on the cluster. However, it can be done through Patroni by adding the [standby_cluster] (https://github.com/zalando/patroni/blob/bd2c54581abb42a7d3a3da551edf0b8732eefd27/docs/replica_bootstrap.rst#standby-cluster) section using patronictl edit-config. Note that the transformed standby cluster will not be doing any streaming, rather will just be in standby mode and allow read-only transactions only.
|
||||||
|
|
||||||
## Sidecar Support
|
## Sidecar Support
|
||||||
|
|
||||||
Each cluster can specify arbitrary sidecars to run. These containers could be used for
|
Each cluster can specify arbitrary sidecars to run. These containers could be used for
|
||||||
|
|
|
||||||
|
|
@ -86,4 +86,3 @@ spec:
|
||||||
# env:
|
# env:
|
||||||
# - name: "USEFUL_VAR"
|
# - name: "USEFUL_VAR"
|
||||||
# value: "perhaps-true"
|
# value: "perhaps-true"
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,20 @@
|
||||||
|
|
||||||
|
apiVersion: "acid.zalan.do/v1"
|
||||||
|
kind: postgresql
|
||||||
|
metadata:
|
||||||
|
name: acid-standby-cluster
|
||||||
|
namespace: default
|
||||||
|
spec:
|
||||||
|
teamId: "ACID"
|
||||||
|
volume:
|
||||||
|
size: 1Gi
|
||||||
|
numberOfInstances: 1
|
||||||
|
postgresql:
|
||||||
|
version: "10"
|
||||||
|
# Make this a standby cluster and provide the s3 bucket path of source cluster for continuous streaming.
|
||||||
|
standby:
|
||||||
|
s3_wal_path: "s3://path/to/bucket/containing/wal/of/source/cluster/"
|
||||||
|
|
||||||
|
maintenanceWindows:
|
||||||
|
- 01:00-06:00 #UTC
|
||||||
|
- Sat:00:00-04:00
|
||||||
|
|
@ -58,6 +58,7 @@ type PostgresSpec struct {
|
||||||
ShmVolume *bool `json:"enableShmVolume,omitempty"`
|
ShmVolume *bool `json:"enableShmVolume,omitempty"`
|
||||||
EnableLogicalBackup bool `json:"enableLogicalBackup,omitempty"`
|
EnableLogicalBackup bool `json:"enableLogicalBackup,omitempty"`
|
||||||
LogicalBackupSchedule string `json:"logicalBackupSchedule,omitempty"`
|
LogicalBackupSchedule string `json:"logicalBackupSchedule,omitempty"`
|
||||||
|
StandbyCluster *StandbyDescription `json:"standby"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
|
||||||
|
|
@ -114,6 +115,11 @@ type Patroni struct {
|
||||||
Slots map[string]map[string]string `json:"slots"`
|
Slots map[string]map[string]string `json:"slots"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//StandbyCluster
|
||||||
|
type StandbyDescription struct {
|
||||||
|
S3WalPath string `json:"s3_wal_path,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
// CloneDescription describes which cluster the new should clone and up to which point in time
|
// CloneDescription describes which cluster the new should clone and up to which point in time
|
||||||
type CloneDescription struct {
|
type CloneDescription struct {
|
||||||
ClusterName string `json:"cluster,omitempty"`
|
ClusterName string `json:"cluster,omitempty"`
|
||||||
|
|
|
||||||
|
|
@ -354,6 +354,28 @@ var unmarshalCluster = []struct {
|
||||||
},
|
},
|
||||||
marshal: []byte(`{"kind":"Postgresql","apiVersion":"acid.zalan.do/v1","metadata":{"name":"acid-testcluster1","creationTimestamp":null},"spec":{"postgresql":{"version":"","parameters":null},"volume":{"size":"","storageClass":""},"patroni":{"initdb":null,"pg_hba":null,"ttl":0,"loop_wait":0,"retry_timeout":0,"maximum_lag_on_failover":0,"slots":null},"resources":{"requests":{"cpu":"","memory":""},"limits":{"cpu":"","memory":""}},"teamId":"acid","allowedSourceRanges":null,"numberOfInstances":0,"users":null,"clone":{"cluster":"team-batman"}},"status":{"PostgresClusterStatus":""}}`),
|
marshal: []byte(`{"kind":"Postgresql","apiVersion":"acid.zalan.do/v1","metadata":{"name":"acid-testcluster1","creationTimestamp":null},"spec":{"postgresql":{"version":"","parameters":null},"volume":{"size":"","storageClass":""},"patroni":{"initdb":null,"pg_hba":null,"ttl":0,"loop_wait":0,"retry_timeout":0,"maximum_lag_on_failover":0,"slots":null},"resources":{"requests":{"cpu":"","memory":""},"limits":{"cpu":"","memory":""}},"teamId":"acid","allowedSourceRanges":null,"numberOfInstances":0,"users":null,"clone":{"cluster":"team-batman"}},"status":{"PostgresClusterStatus":""}}`),
|
||||||
err: nil},
|
err: nil},
|
||||||
|
// standby example
|
||||||
|
{
|
||||||
|
in: []byte(`{"kind": "Postgresql","apiVersion": "acid.zalan.do/v1","metadata": {"name": "acid-testcluster1"}, "spec": {"teamId": "acid", "standby": {"s3_wal_path": "s3://custom/path/to/bucket/"}}}`),
|
||||||
|
out: Postgresql{
|
||||||
|
TypeMeta: metav1.TypeMeta{
|
||||||
|
Kind: "Postgresql",
|
||||||
|
APIVersion: "acid.zalan.do/v1",
|
||||||
|
},
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: "acid-testcluster1",
|
||||||
|
},
|
||||||
|
Spec: PostgresSpec{
|
||||||
|
TeamID: "acid",
|
||||||
|
StandbyCluster: &StandbyDescription{
|
||||||
|
S3WalPath: "s3://custom/path/to/bucket/",
|
||||||
|
},
|
||||||
|
ClusterName: "testcluster1",
|
||||||
|
},
|
||||||
|
Error: "",
|
||||||
|
},
|
||||||
|
marshal: []byte(`{"kind":"Postgresql","apiVersion":"acid.zalan.do/v1","metadata":{"name":"acid-testcluster1","creationTimestamp":null},"spec":{"postgresql":{"version":"","parameters":null},"volume":{"size":"","storageClass":""},"patroni":{"initdb":null,"pg_hba":null,"ttl":0,"loop_wait":0,"retry_timeout":0,"maximum_lag_on_failover":0,"slots":null},"resources":{"requests":{"cpu":"","memory":""},"limits":{"cpu":"","memory":""}},"teamId":"acid","allowedSourceRanges":null,"numberOfInstances":0,"users":null,"standby":{"s3_wal_path":"s3://custom/path/to/bucket/"}},"status":{"PostgresClusterStatus":""}}`),
|
||||||
|
err: nil},
|
||||||
// erroneous examples
|
// erroneous examples
|
||||||
{
|
{
|
||||||
in: []byte(`{"kind": "Postgresql","apiVersion": "acid.zalan.do/v1"`),
|
in: []byte(`{"kind": "Postgresql","apiVersion": "acid.zalan.do/v1"`),
|
||||||
|
|
|
||||||
|
|
@ -503,6 +503,11 @@ func (in *PostgresSpec) DeepCopyInto(out *PostgresSpec) {
|
||||||
*out = new(bool)
|
*out = new(bool)
|
||||||
**out = **in
|
**out = **in
|
||||||
}
|
}
|
||||||
|
if in.StandbyCluster != nil {
|
||||||
|
in, out := &in.StandbyCluster, &out.StandbyCluster
|
||||||
|
*out = new(StandbyDescription)
|
||||||
|
**out = **in
|
||||||
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -711,6 +716,22 @@ func (in *Sidecar) DeepCopy() *Sidecar {
|
||||||
return out
|
return out
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||||
|
func (in *StandbyDescription) DeepCopyInto(out *StandbyDescription) {
|
||||||
|
*out = *in
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new StandbyDescription.
|
||||||
|
func (in *StandbyDescription) DeepCopy() *StandbyDescription {
|
||||||
|
if in == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
out := new(StandbyDescription)
|
||||||
|
in.DeepCopyInto(out)
|
||||||
|
return out
|
||||||
|
}
|
||||||
|
|
||||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||||
func (in *TeamsAPIConfiguration) DeepCopyInto(out *TeamsAPIConfiguration) {
|
func (in *TeamsAPIConfiguration) DeepCopyInto(out *TeamsAPIConfiguration) {
|
||||||
*out = *in
|
*out = *in
|
||||||
|
|
|
||||||
|
|
@ -287,7 +287,7 @@ func (c *Cluster) Create() error {
|
||||||
c.logger.Infof("pods are ready")
|
c.logger.Infof("pods are ready")
|
||||||
|
|
||||||
// create database objects unless we are running without pods or disabled that feature explicitly
|
// create database objects unless we are running without pods or disabled that feature explicitly
|
||||||
if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&c.Spec) <= 0) {
|
if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&c.Spec) <= 0 || c.Spec.StandbyCluster != nil) {
|
||||||
if err = c.createRoles(); err != nil {
|
if err = c.createRoles(); err != nil {
|
||||||
return fmt.Errorf("could not create users: %v", err)
|
return fmt.Errorf("could not create users: %v", err)
|
||||||
}
|
}
|
||||||
|
|
@ -626,7 +626,7 @@ func (c *Cluster) Update(oldSpec, newSpec *acidv1.Postgresql) error {
|
||||||
}()
|
}()
|
||||||
|
|
||||||
// Roles and Databases
|
// Roles and Databases
|
||||||
if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&c.Spec) <= 0) {
|
if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&c.Spec) <= 0 || c.Spec.StandbyCluster != nil) {
|
||||||
c.logger.Debugf("syncing roles")
|
c.logger.Debugf("syncing roles")
|
||||||
if err := c.syncRoles(); err != nil {
|
if err := c.syncRoles(); err != nil {
|
||||||
c.logger.Errorf("could not sync roles: %v", err)
|
c.logger.Errorf("could not sync roles: %v", err)
|
||||||
|
|
|
||||||
|
|
@ -500,7 +500,7 @@ func generatePodTemplate(
|
||||||
}
|
}
|
||||||
|
|
||||||
// generatePodEnvVars generates environment variables for the Spilo Pod
|
// generatePodEnvVars generates environment variables for the Spilo Pod
|
||||||
func (c *Cluster) generateSpiloPodEnvVars(uid types.UID, spiloConfiguration string, cloneDescription *acidv1.CloneDescription, customPodEnvVarsList []v1.EnvVar) []v1.EnvVar {
|
func (c *Cluster) generateSpiloPodEnvVars(uid types.UID, spiloConfiguration string, cloneDescription *acidv1.CloneDescription, standbyDescription *acidv1.StandbyDescription, customPodEnvVarsList []v1.EnvVar) []v1.EnvVar {
|
||||||
envVars := []v1.EnvVar{
|
envVars := []v1.EnvVar{
|
||||||
{
|
{
|
||||||
Name: "SCOPE",
|
Name: "SCOPE",
|
||||||
|
|
@ -604,6 +604,10 @@ func (c *Cluster) generateSpiloPodEnvVars(uid types.UID, spiloConfiguration stri
|
||||||
envVars = append(envVars, c.generateCloneEnvironment(cloneDescription)...)
|
envVars = append(envVars, c.generateCloneEnvironment(cloneDescription)...)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if c.Spec.StandbyCluster != nil {
|
||||||
|
envVars = append(envVars, c.generateStandbyEnvironment(standbyDescription)...)
|
||||||
|
}
|
||||||
|
|
||||||
if len(customPodEnvVarsList) > 0 {
|
if len(customPodEnvVarsList) > 0 {
|
||||||
envVars = append(envVars, customPodEnvVarsList...)
|
envVars = append(envVars, customPodEnvVarsList...)
|
||||||
}
|
}
|
||||||
|
|
@ -793,6 +797,9 @@ func (c *Cluster) generateStatefulSet(spec *acidv1.PostgresSpec) (*v1beta1.State
|
||||||
sort.Slice(customPodEnvVarsList,
|
sort.Slice(customPodEnvVarsList,
|
||||||
func(i, j int) bool { return customPodEnvVarsList[i].Name < customPodEnvVarsList[j].Name })
|
func(i, j int) bool { return customPodEnvVarsList[i].Name < customPodEnvVarsList[j].Name })
|
||||||
}
|
}
|
||||||
|
if spec.StandbyCluster != nil && spec.StandbyCluster.S3WalPath == "" {
|
||||||
|
return nil, fmt.Errorf("s3_wal_path is empty for standby cluster")
|
||||||
|
}
|
||||||
|
|
||||||
spiloConfiguration, err := generateSpiloJSONConfiguration(&spec.PostgresqlParam, &spec.Patroni, c.OpConfig.PamRoleName, c.logger)
|
spiloConfiguration, err := generateSpiloJSONConfiguration(&spec.PostgresqlParam, &spec.Patroni, c.OpConfig.PamRoleName, c.logger)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
@ -802,7 +809,7 @@ func (c *Cluster) generateStatefulSet(spec *acidv1.PostgresSpec) (*v1beta1.State
|
||||||
// generate environment variables for the spilo container
|
// generate environment variables for the spilo container
|
||||||
spiloEnvVars := deduplicateEnvVars(
|
spiloEnvVars := deduplicateEnvVars(
|
||||||
c.generateSpiloPodEnvVars(c.Postgresql.GetUID(), spiloConfiguration, &spec.Clone,
|
c.generateSpiloPodEnvVars(c.Postgresql.GetUID(), spiloConfiguration, &spec.Clone,
|
||||||
customPodEnvVarsList), c.containerName(), c.logger)
|
spec.StandbyCluster, customPodEnvVarsList), c.containerName(), c.logger)
|
||||||
|
|
||||||
// pickup the docker image for the spilo container
|
// pickup the docker image for the spilo container
|
||||||
effectiveDockerImage := util.Coalesce(spec.DockerImage, c.OpConfig.DockerImage)
|
effectiveDockerImage := util.Coalesce(spec.DockerImage, c.OpConfig.DockerImage)
|
||||||
|
|
@ -982,6 +989,11 @@ func (c *Cluster) getNumberOfInstances(spec *acidv1.PostgresSpec) int32 {
|
||||||
cur := spec.NumberOfInstances
|
cur := spec.NumberOfInstances
|
||||||
newcur := cur
|
newcur := cur
|
||||||
|
|
||||||
|
/* Limit the max number of pods to one, if this is standby-cluster */
|
||||||
|
if spec.StandbyCluster != nil {
|
||||||
|
c.logger.Info("Standby cluster can have maximum of 1 pod")
|
||||||
|
max = 1
|
||||||
|
}
|
||||||
if max >= 0 && newcur > max {
|
if max >= 0 && newcur > max {
|
||||||
newcur = max
|
newcur = max
|
||||||
}
|
}
|
||||||
|
|
@ -1328,6 +1340,27 @@ func (c *Cluster) generateCloneEnvironment(description *acidv1.CloneDescription)
|
||||||
return result
|
return result
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *Cluster) generateStandbyEnvironment(description *acidv1.StandbyDescription) []v1.EnvVar {
|
||||||
|
result := make([]v1.EnvVar, 0)
|
||||||
|
|
||||||
|
if description.S3WalPath == "" {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
// standby with S3, find out the bucket to setup standby
|
||||||
|
msg := "Standby from S3 bucket using custom parsed S3WalPath from the manifest %s "
|
||||||
|
c.logger.Infof(msg, description.S3WalPath)
|
||||||
|
|
||||||
|
result = append(result, v1.EnvVar{
|
||||||
|
Name: "STANDBY_WALE_S3_PREFIX",
|
||||||
|
Value: description.S3WalPath,
|
||||||
|
})
|
||||||
|
|
||||||
|
result = append(result, v1.EnvVar{Name: "STANDBY_METHOD", Value: "STANDBY_WITH_WALE"})
|
||||||
|
result = append(result, v1.EnvVar{Name: "STANDBY_WAL_BUCKET_SCOPE_PREFIX", Value: ""})
|
||||||
|
|
||||||
|
return result
|
||||||
|
}
|
||||||
|
|
||||||
func (c *Cluster) generatePodDisruptionBudget() *policybeta1.PodDisruptionBudget {
|
func (c *Cluster) generatePodDisruptionBudget() *policybeta1.PodDisruptionBudget {
|
||||||
minAvailable := intstr.FromInt(1)
|
minAvailable := intstr.FromInt(1)
|
||||||
pdbEnabled := c.OpConfig.EnablePodDisruptionBudget
|
pdbEnabled := c.OpConfig.EnablePodDisruptionBudget
|
||||||
|
|
|
||||||
|
|
@ -90,7 +90,7 @@ func (c *Cluster) Sync(newSpec *acidv1.Postgresql) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// create database objects unless we are running without pods or disabled that feature explicitly
|
// create database objects unless we are running without pods or disabled that feature explicitly
|
||||||
if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&newSpec.Spec) <= 0) {
|
if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&newSpec.Spec) <= 0 || c.Spec.StandbyCluster != nil) {
|
||||||
c.logger.Debugf("syncing roles")
|
c.logger.Debugf("syncing roles")
|
||||||
if err = c.syncRoles(); err != nil {
|
if err = c.syncRoles(); err != nil {
|
||||||
err = fmt.Errorf("could not sync roles: %v", err)
|
err = fmt.Errorf("could not sync roles: %v", err)
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue