463 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			463 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			Go
		
	
	
	
package cluster
 | 
						|
 | 
						|
import (
 | 
						|
	"reflect"
 | 
						|
 | 
						|
	"k8s.io/api/core/v1"
 | 
						|
 | 
						|
	"testing"
 | 
						|
 | 
						|
	acidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1"
 | 
						|
	"github.com/zalando/postgres-operator/pkg/util/config"
 | 
						|
	"github.com/zalando/postgres-operator/pkg/util/constants"
 | 
						|
	"github.com/zalando/postgres-operator/pkg/util/k8sutil"
 | 
						|
 | 
						|
	policyv1beta1 "k8s.io/api/policy/v1beta1"
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/util/intstr"
 | 
						|
)
 | 
						|
 | 
						|
func True() *bool {
 | 
						|
	b := true
 | 
						|
	return &b
 | 
						|
}
 | 
						|
 | 
						|
func False() *bool {
 | 
						|
	b := false
 | 
						|
	return &b
 | 
						|
}
 | 
						|
 | 
						|
func toIntStr(val int) *intstr.IntOrString {
 | 
						|
	b := intstr.FromInt(val)
 | 
						|
	return &b
 | 
						|
}
 | 
						|
 | 
						|
func TestGenerateSpiloJSONConfiguration(t *testing.T) {
 | 
						|
	var cluster = New(
 | 
						|
		Config{
 | 
						|
			OpConfig: config.Config{
 | 
						|
				ProtectedRoles: []string{"admin"},
 | 
						|
				Auth: config.Auth{
 | 
						|
					SuperUsername:       superUserName,
 | 
						|
					ReplicationUsername: replicationUserName,
 | 
						|
				},
 | 
						|
			},
 | 
						|
		}, k8sutil.KubernetesClient{}, acidv1.Postgresql{}, logger)
 | 
						|
 | 
						|
	testName := "TestGenerateSpiloConfig"
 | 
						|
	tests := []struct {
 | 
						|
		subtest  string
 | 
						|
		pgParam  *acidv1.PostgresqlParam
 | 
						|
		patroni  *acidv1.Patroni
 | 
						|
		role     string
 | 
						|
		opConfig config.Config
 | 
						|
		result   string
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			subtest:  "Patroni default configuration",
 | 
						|
			pgParam:  &acidv1.PostgresqlParam{PgVersion: "9.6"},
 | 
						|
			patroni:  &acidv1.Patroni{},
 | 
						|
			role:     "zalandos",
 | 
						|
			opConfig: config.Config{},
 | 
						|
			result:   `{"postgresql":{"bin_dir":"/usr/lib/postgresql/9.6/bin"},"bootstrap":{"initdb":[{"auth-host":"md5"},{"auth-local":"trust"}],"users":{"zalandos":{"password":"","options":["CREATEDB","NOLOGIN"]}},"dcs":{}}}`,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subtest: "Patroni configured",
 | 
						|
			pgParam: &acidv1.PostgresqlParam{PgVersion: "11"},
 | 
						|
			patroni: &acidv1.Patroni{
 | 
						|
				InitDB: map[string]string{
 | 
						|
					"encoding":       "UTF8",
 | 
						|
					"locale":         "en_US.UTF-8",
 | 
						|
					"data-checksums": "true",
 | 
						|
				},
 | 
						|
				PgHba:                []string{"hostssl all all 0.0.0.0/0 md5", "host    all all 0.0.0.0/0 md5"},
 | 
						|
				TTL:                  30,
 | 
						|
				LoopWait:             10,
 | 
						|
				RetryTimeout:         10,
 | 
						|
				MaximumLagOnFailover: 33554432,
 | 
						|
				Slots:                map[string]map[string]string{"permanent_logical_1": {"type": "logical", "database": "foo", "plugin": "pgoutput"}},
 | 
						|
			},
 | 
						|
			role:     "zalandos",
 | 
						|
			opConfig: config.Config{},
 | 
						|
			result:   `{"postgresql":{"bin_dir":"/usr/lib/postgresql/11/bin","pg_hba":["hostssl all all 0.0.0.0/0 md5","host    all all 0.0.0.0/0 md5"]},"bootstrap":{"initdb":[{"auth-host":"md5"},{"auth-local":"trust"},"data-checksums",{"encoding":"UTF8"},{"locale":"en_US.UTF-8"}],"users":{"zalandos":{"password":"","options":["CREATEDB","NOLOGIN"]}},"dcs":{"ttl":30,"loop_wait":10,"retry_timeout":10,"maximum_lag_on_failover":33554432,"slots":{"permanent_logical_1":{"database":"foo","plugin":"pgoutput","type":"logical"}}}}}`,
 | 
						|
		},
 | 
						|
	}
 | 
						|
	for _, tt := range tests {
 | 
						|
		cluster.OpConfig = tt.opConfig
 | 
						|
		result, err := generateSpiloJSONConfiguration(tt.pgParam, tt.patroni, tt.role, logger)
 | 
						|
		if err != nil {
 | 
						|
			t.Errorf("Unexpected error: %v", err)
 | 
						|
		}
 | 
						|
		if tt.result != result {
 | 
						|
			t.Errorf("%s %s: Spilo Config is %v, expected %v for role %#v and param %#v",
 | 
						|
				testName, tt.subtest, result, tt.result, tt.role, tt.pgParam)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestCreateLoadBalancerLogic(t *testing.T) {
 | 
						|
	var cluster = New(
 | 
						|
		Config{
 | 
						|
			OpConfig: config.Config{
 | 
						|
				ProtectedRoles: []string{"admin"},
 | 
						|
				Auth: config.Auth{
 | 
						|
					SuperUsername:       superUserName,
 | 
						|
					ReplicationUsername: replicationUserName,
 | 
						|
				},
 | 
						|
			},
 | 
						|
		}, k8sutil.KubernetesClient{}, acidv1.Postgresql{}, logger)
 | 
						|
 | 
						|
	testName := "TestCreateLoadBalancerLogic"
 | 
						|
	tests := []struct {
 | 
						|
		subtest  string
 | 
						|
		role     PostgresRole
 | 
						|
		spec     *acidv1.PostgresSpec
 | 
						|
		opConfig config.Config
 | 
						|
		result   bool
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			subtest:  "new format, load balancer is enabled for replica",
 | 
						|
			role:     Replica,
 | 
						|
			spec:     &acidv1.PostgresSpec{EnableReplicaLoadBalancer: True()},
 | 
						|
			opConfig: config.Config{},
 | 
						|
			result:   true,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subtest:  "new format, load balancer is disabled for replica",
 | 
						|
			role:     Replica,
 | 
						|
			spec:     &acidv1.PostgresSpec{EnableReplicaLoadBalancer: False()},
 | 
						|
			opConfig: config.Config{},
 | 
						|
			result:   false,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subtest:  "new format, load balancer isn't specified for replica",
 | 
						|
			role:     Replica,
 | 
						|
			spec:     &acidv1.PostgresSpec{EnableReplicaLoadBalancer: nil},
 | 
						|
			opConfig: config.Config{EnableReplicaLoadBalancer: true},
 | 
						|
			result:   true,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subtest:  "new format, load balancer isn't specified for replica",
 | 
						|
			role:     Replica,
 | 
						|
			spec:     &acidv1.PostgresSpec{EnableReplicaLoadBalancer: nil},
 | 
						|
			opConfig: config.Config{EnableReplicaLoadBalancer: false},
 | 
						|
			result:   false,
 | 
						|
		},
 | 
						|
	}
 | 
						|
	for _, tt := range tests {
 | 
						|
		cluster.OpConfig = tt.opConfig
 | 
						|
		result := cluster.shouldCreateLoadBalancerForService(tt.role, tt.spec)
 | 
						|
		if tt.result != result {
 | 
						|
			t.Errorf("%s %s: Load balancer is %t, expect %t for role %#v and spec %#v",
 | 
						|
				testName, tt.subtest, result, tt.result, tt.role, tt.spec)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestGeneratePodDisruptionBudget(t *testing.T) {
 | 
						|
	tests := []struct {
 | 
						|
		c   *Cluster
 | 
						|
		out policyv1beta1.PodDisruptionBudget
 | 
						|
	}{
 | 
						|
		// With multiple instances.
 | 
						|
		{
 | 
						|
			New(
 | 
						|
				Config{OpConfig: config.Config{Resources: config.Resources{ClusterNameLabel: "cluster-name", PodRoleLabel: "spilo-role"}, PDBNameFormat: "postgres-{cluster}-pdb"}},
 | 
						|
				k8sutil.KubernetesClient{},
 | 
						|
				acidv1.Postgresql{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: "myapp-database", Namespace: "myapp"},
 | 
						|
					Spec:       acidv1.PostgresSpec{TeamID: "myapp", NumberOfInstances: 3}},
 | 
						|
				logger),
 | 
						|
			policyv1beta1.PodDisruptionBudget{
 | 
						|
				ObjectMeta: metav1.ObjectMeta{
 | 
						|
					Name:      "postgres-myapp-database-pdb",
 | 
						|
					Namespace: "myapp",
 | 
						|
					Labels:    map[string]string{"team": "myapp", "cluster-name": "myapp-database"},
 | 
						|
				},
 | 
						|
				Spec: policyv1beta1.PodDisruptionBudgetSpec{
 | 
						|
					MinAvailable: toIntStr(1),
 | 
						|
					Selector: &metav1.LabelSelector{
 | 
						|
						MatchLabels: map[string]string{"spilo-role": "master", "cluster-name": "myapp-database"},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
		},
 | 
						|
		// With zero instances.
 | 
						|
		{
 | 
						|
			New(
 | 
						|
				Config{OpConfig: config.Config{Resources: config.Resources{ClusterNameLabel: "cluster-name", PodRoleLabel: "spilo-role"}, PDBNameFormat: "postgres-{cluster}-pdb"}},
 | 
						|
				k8sutil.KubernetesClient{},
 | 
						|
				acidv1.Postgresql{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: "myapp-database", Namespace: "myapp"},
 | 
						|
					Spec:       acidv1.PostgresSpec{TeamID: "myapp", NumberOfInstances: 0}},
 | 
						|
				logger),
 | 
						|
			policyv1beta1.PodDisruptionBudget{
 | 
						|
				ObjectMeta: metav1.ObjectMeta{
 | 
						|
					Name:      "postgres-myapp-database-pdb",
 | 
						|
					Namespace: "myapp",
 | 
						|
					Labels:    map[string]string{"team": "myapp", "cluster-name": "myapp-database"},
 | 
						|
				},
 | 
						|
				Spec: policyv1beta1.PodDisruptionBudgetSpec{
 | 
						|
					MinAvailable: toIntStr(0),
 | 
						|
					Selector: &metav1.LabelSelector{
 | 
						|
						MatchLabels: map[string]string{"spilo-role": "master", "cluster-name": "myapp-database"},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
		},
 | 
						|
		// With PodDisruptionBudget disabled.
 | 
						|
		{
 | 
						|
			New(
 | 
						|
				Config{OpConfig: config.Config{Resources: config.Resources{ClusterNameLabel: "cluster-name", PodRoleLabel: "spilo-role"}, PDBNameFormat: "postgres-{cluster}-pdb", EnablePodDisruptionBudget: False()}},
 | 
						|
				k8sutil.KubernetesClient{},
 | 
						|
				acidv1.Postgresql{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: "myapp-database", Namespace: "myapp"},
 | 
						|
					Spec:       acidv1.PostgresSpec{TeamID: "myapp", NumberOfInstances: 3}},
 | 
						|
				logger),
 | 
						|
			policyv1beta1.PodDisruptionBudget{
 | 
						|
				ObjectMeta: metav1.ObjectMeta{
 | 
						|
					Name:      "postgres-myapp-database-pdb",
 | 
						|
					Namespace: "myapp",
 | 
						|
					Labels:    map[string]string{"team": "myapp", "cluster-name": "myapp-database"},
 | 
						|
				},
 | 
						|
				Spec: policyv1beta1.PodDisruptionBudgetSpec{
 | 
						|
					MinAvailable: toIntStr(0),
 | 
						|
					Selector: &metav1.LabelSelector{
 | 
						|
						MatchLabels: map[string]string{"spilo-role": "master", "cluster-name": "myapp-database"},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
		},
 | 
						|
		// With non-default PDBNameFormat and PodDisruptionBudget explicitly enabled.
 | 
						|
		{
 | 
						|
			New(
 | 
						|
				Config{OpConfig: config.Config{Resources: config.Resources{ClusterNameLabel: "cluster-name", PodRoleLabel: "spilo-role"}, PDBNameFormat: "postgres-{cluster}-databass-budget", EnablePodDisruptionBudget: True()}},
 | 
						|
				k8sutil.KubernetesClient{},
 | 
						|
				acidv1.Postgresql{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: "myapp-database", Namespace: "myapp"},
 | 
						|
					Spec:       acidv1.PostgresSpec{TeamID: "myapp", NumberOfInstances: 3}},
 | 
						|
				logger),
 | 
						|
			policyv1beta1.PodDisruptionBudget{
 | 
						|
				ObjectMeta: metav1.ObjectMeta{
 | 
						|
					Name:      "postgres-myapp-database-databass-budget",
 | 
						|
					Namespace: "myapp",
 | 
						|
					Labels:    map[string]string{"team": "myapp", "cluster-name": "myapp-database"},
 | 
						|
				},
 | 
						|
				Spec: policyv1beta1.PodDisruptionBudgetSpec{
 | 
						|
					MinAvailable: toIntStr(1),
 | 
						|
					Selector: &metav1.LabelSelector{
 | 
						|
						MatchLabels: map[string]string{"spilo-role": "master", "cluster-name": "myapp-database"},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for _, tt := range tests {
 | 
						|
		result := tt.c.generatePodDisruptionBudget()
 | 
						|
		if !reflect.DeepEqual(*result, tt.out) {
 | 
						|
			t.Errorf("Expected PodDisruptionBudget: %#v, got %#v", tt.out, *result)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestShmVolume(t *testing.T) {
 | 
						|
	testName := "TestShmVolume"
 | 
						|
	tests := []struct {
 | 
						|
		subTest string
 | 
						|
		podSpec *v1.PodSpec
 | 
						|
		shmPos  int
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			subTest: "empty PodSpec",
 | 
						|
			podSpec: &v1.PodSpec{
 | 
						|
				Volumes: []v1.Volume{},
 | 
						|
				Containers: []v1.Container{
 | 
						|
					{
 | 
						|
						VolumeMounts: []v1.VolumeMount{},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
			shmPos: 0,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subTest: "non empty PodSpec",
 | 
						|
			podSpec: &v1.PodSpec{
 | 
						|
				Volumes: []v1.Volume{{}},
 | 
						|
				Containers: []v1.Container{
 | 
						|
					{
 | 
						|
						VolumeMounts: []v1.VolumeMount{
 | 
						|
							{},
 | 
						|
						},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
			shmPos: 1,
 | 
						|
		},
 | 
						|
	}
 | 
						|
	for _, tt := range tests {
 | 
						|
		addShmVolume(tt.podSpec)
 | 
						|
 | 
						|
		volumeName := tt.podSpec.Volumes[tt.shmPos].Name
 | 
						|
		volumeMountName := tt.podSpec.Containers[0].VolumeMounts[tt.shmPos].Name
 | 
						|
 | 
						|
		if volumeName != constants.ShmVolumeName {
 | 
						|
			t.Errorf("%s %s: Expected volume %s was not created, have %s instead",
 | 
						|
				testName, tt.subTest, constants.ShmVolumeName, volumeName)
 | 
						|
		}
 | 
						|
		if volumeMountName != constants.ShmVolumeName {
 | 
						|
			t.Errorf("%s %s: Expected mount %s was not created, have %s instead",
 | 
						|
				testName, tt.subTest, constants.ShmVolumeName, volumeMountName)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestCloneEnv(t *testing.T) {
 | 
						|
	testName := "TestCloneEnv"
 | 
						|
	tests := []struct {
 | 
						|
		subTest   string
 | 
						|
		cloneOpts *acidv1.CloneDescription
 | 
						|
		env       v1.EnvVar
 | 
						|
		envPos    int
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			subTest: "custom s3 path",
 | 
						|
			cloneOpts: &acidv1.CloneDescription{
 | 
						|
				ClusterName:  "test-cluster",
 | 
						|
				S3WalPath:    "s3://some/path/",
 | 
						|
				EndTimestamp: "somewhen",
 | 
						|
			},
 | 
						|
			env: v1.EnvVar{
 | 
						|
				Name:  "CLONE_WALE_S3_PREFIX",
 | 
						|
				Value: "s3://some/path/",
 | 
						|
			},
 | 
						|
			envPos: 1,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subTest: "generated s3 path, bucket",
 | 
						|
			cloneOpts: &acidv1.CloneDescription{
 | 
						|
				ClusterName:  "test-cluster",
 | 
						|
				EndTimestamp: "somewhen",
 | 
						|
				UID:          "0000",
 | 
						|
			},
 | 
						|
			env: v1.EnvVar{
 | 
						|
				Name:  "CLONE_WAL_S3_BUCKET",
 | 
						|
				Value: "wale-bucket",
 | 
						|
			},
 | 
						|
			envPos: 1,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subTest: "generated s3 path, target time",
 | 
						|
			cloneOpts: &acidv1.CloneDescription{
 | 
						|
				ClusterName:  "test-cluster",
 | 
						|
				EndTimestamp: "somewhen",
 | 
						|
				UID:          "0000",
 | 
						|
			},
 | 
						|
			env: v1.EnvVar{
 | 
						|
				Name:  "CLONE_TARGET_TIME",
 | 
						|
				Value: "somewhen",
 | 
						|
			},
 | 
						|
			envPos: 4,
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	var cluster = New(
 | 
						|
		Config{
 | 
						|
			OpConfig: config.Config{
 | 
						|
				WALES3Bucket:   "wale-bucket",
 | 
						|
				ProtectedRoles: []string{"admin"},
 | 
						|
				Auth: config.Auth{
 | 
						|
					SuperUsername:       superUserName,
 | 
						|
					ReplicationUsername: replicationUserName,
 | 
						|
				},
 | 
						|
			},
 | 
						|
		}, k8sutil.KubernetesClient{}, acidv1.Postgresql{}, logger)
 | 
						|
 | 
						|
	for _, tt := range tests {
 | 
						|
		envs := cluster.generateCloneEnvironment(tt.cloneOpts)
 | 
						|
 | 
						|
		env := envs[tt.envPos]
 | 
						|
 | 
						|
		if env.Name != tt.env.Name {
 | 
						|
			t.Errorf("%s %s: Expected env name %s, have %s instead",
 | 
						|
				testName, tt.subTest, tt.env.Name, env.Name)
 | 
						|
		}
 | 
						|
 | 
						|
		if env.Value != tt.env.Value {
 | 
						|
			t.Errorf("%s %s: Expected env value %s, have %s instead",
 | 
						|
				testName, tt.subTest, tt.env.Value, env.Value)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestSecretVolume(t *testing.T) {
 | 
						|
	testName := "TestSecretVolume"
 | 
						|
	tests := []struct {
 | 
						|
		subTest   string
 | 
						|
		podSpec   *v1.PodSpec
 | 
						|
		secretPos int
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			subTest: "empty PodSpec",
 | 
						|
			podSpec: &v1.PodSpec{
 | 
						|
				Volumes: []v1.Volume{},
 | 
						|
				Containers: []v1.Container{
 | 
						|
					{
 | 
						|
						VolumeMounts: []v1.VolumeMount{},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
			secretPos: 0,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			subTest: "non empty PodSpec",
 | 
						|
			podSpec: &v1.PodSpec{
 | 
						|
				Volumes: []v1.Volume{{}},
 | 
						|
				Containers: []v1.Container{
 | 
						|
					{
 | 
						|
						VolumeMounts: []v1.VolumeMount{
 | 
						|
							{
 | 
						|
								Name:      "data",
 | 
						|
								ReadOnly:  false,
 | 
						|
								MountPath: "/data",
 | 
						|
							},
 | 
						|
						},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
			secretPos: 1,
 | 
						|
		},
 | 
						|
	}
 | 
						|
	for _, tt := range tests {
 | 
						|
		additionalSecretMount := "aws-iam-s3-role"
 | 
						|
		additionalSecretMountPath := "/meta/credentials"
 | 
						|
 | 
						|
		numMounts := len(tt.podSpec.Containers[0].VolumeMounts)
 | 
						|
 | 
						|
		addSecretVolume(tt.podSpec, additionalSecretMount, additionalSecretMountPath)
 | 
						|
 | 
						|
		volumeName := tt.podSpec.Volumes[tt.secretPos].Name
 | 
						|
 | 
						|
		if volumeName != additionalSecretMount {
 | 
						|
			t.Errorf("%s %s: Expected volume %s was not created, have %s instead",
 | 
						|
				testName, tt.subTest, additionalSecretMount, volumeName)
 | 
						|
		}
 | 
						|
 | 
						|
		for i := range tt.podSpec.Containers {
 | 
						|
			volumeMountName := tt.podSpec.Containers[i].VolumeMounts[tt.secretPos].Name
 | 
						|
 | 
						|
			if volumeMountName != additionalSecretMount {
 | 
						|
				t.Errorf("%s %s: Expected mount %s was not created, have %s instead",
 | 
						|
					testName, tt.subTest, additionalSecretMount, volumeMountName)
 | 
						|
			}
 | 
						|
		}
 | 
						|
 | 
						|
		numMountsCheck := len(tt.podSpec.Containers[0].VolumeMounts)
 | 
						|
 | 
						|
		if numMountsCheck != numMounts+1 {
 | 
						|
			t.Errorf("Unexpected number of VolumeMounts: got %v instead of %v",
 | 
						|
				numMountsCheck, numMounts+1)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 |