Sync clusters states

* move statefulset creation from cluster spec to the separate function
* sync cluster state with desired state;
* move out from arrays for cluster resources;
* recreate pods instead of deleting them in case of statefulset change
* check for master while creating cluster/updating pods
* simplify retryutil
* list pvc while listing resources
* name kubernetes resources with capital letter
* do rolling update in case of env variables change
This commit is contained in:
Murat Kabilov 2017-03-13 11:33:52 +01:00
parent 1377724b2e
commit 6f7399b36f
13 changed files with 334 additions and 88 deletions

View File

@ -109,6 +109,11 @@ func (c *Cluster) needsRollingUpdate(otherSpec *spec.Postgresql) bool {
return true return true
} }
newSs := genStatefulSet(c.ClusterName(), otherSpec.Spec, c.etcdHost, c.dockerImage)
diff, res := statefulsetsEqual(c.Statefulset, newSs)
if diff && res {
return true
}
return false return false
} }
@ -159,7 +164,7 @@ func (c *Cluster) Create() error {
} }
if err := c.applySecrets(); err != nil { if err := c.applySecrets(); err != nil {
return fmt.Errorf("Can't create secrets: %s", err) return fmt.Errorf("Can't create Secrets: %s", err)
} else { } else {
c.logger.Infof("Secrets have been successfully created") c.logger.Infof("Secrets have been successfully created")
} }
@ -172,8 +177,8 @@ func (c *Cluster) Create() error {
} }
c.logger.Info("Waiting for cluster being ready") c.logger.Info("Waiting for cluster being ready")
err = c.waitStatefulsetPodsReady()
if err != nil { if err := c.waitStatefulsetPodsReady(); err != nil {
c.logger.Errorf("Failed to create cluster: %s", err) c.logger.Errorf("Failed to create cluster: %s", err)
return err return err
} }
@ -202,7 +207,7 @@ func (c *Cluster) Update(newSpec *spec.Postgresql) error {
c.logger.Infof("Pods need to be recreated") c.logger.Infof("Pods need to be recreated")
} }
newStatefulSet := getStatefulSet(c.ClusterName(), newSpec.Spec, c.etcdHost, c.dockerImage) newStatefulSet := genStatefulSet(c.ClusterName(), newSpec.Spec, c.etcdHost, c.dockerImage)
newService := resources.Service(c.ClusterName(), c.TeamName(), newSpec.Spec.AllowedSourceRanges) newService := resources.Service(c.ClusterName(), c.TeamName(), newSpec.Spec.AllowedSourceRanges)
if !servicesEqual(newService, c.Service) { if !servicesEqual(newService, c.Service) {
@ -234,22 +239,25 @@ func (c *Cluster) Update(newSpec *spec.Postgresql) error {
} }
func (c *Cluster) Delete() error { func (c *Cluster) Delete() error {
epName := util.NameFromMeta(c.Endpoint.ObjectMeta)
if err := c.deleteEndpoint(); err != nil { if err := c.deleteEndpoint(); err != nil {
c.logger.Errorf("Can't delete Endpoint: %s", err) c.logger.Errorf("Can't delete Endpoint: %s", err)
} else { } else {
c.logger.Infof("Endpoint '%s' has been deleted", util.NameFromMeta(c.Endpoint.ObjectMeta)) c.logger.Infof("Endpoint '%s' has been deleted", epName)
} }
svcName := util.NameFromMeta(c.Service.ObjectMeta)
if err := c.deleteService(); err != nil { if err := c.deleteService(); err != nil {
c.logger.Errorf("Can't delete Service: %s", err) c.logger.Errorf("Can't delete Service: %s", err)
} else { } else {
c.logger.Infof("Service '%s' has been deleted", util.NameFromMeta(c.Service.ObjectMeta)) c.logger.Infof("Service '%s' has been deleted", svcName)
} }
ssName := util.NameFromMeta(c.Statefulset.ObjectMeta)
if err := c.deleteStatefulSet(); err != nil { if err := c.deleteStatefulSet(); err != nil {
c.logger.Errorf("Can't delete StatefulSet: %s", err) c.logger.Errorf("Can't delete StatefulSet: %s", err)
} else { } else {
c.logger.Infof("StatefulSet '%s' has been deleted", util.NameFromMeta(c.Statefulset.ObjectMeta)) c.logger.Infof("StatefulSet '%s' has been deleted", ssName)
} }
for _, obj := range c.Secrets { for _, obj := range c.Secrets {
@ -265,8 +273,8 @@ func (c *Cluster) Delete() error {
} else { } else {
c.logger.Infof("Pods have been deleted") c.logger.Infof("Pods have been deleted")
} }
err := c.deletePersistenVolumeClaims()
if err != nil { if err := c.deletePersistenVolumeClaims(); err != nil {
return fmt.Errorf("Can't delete PersistentVolumeClaims: %s", err) return fmt.Errorf("Can't delete PersistentVolumeClaims: %s", err)
} }

View File

@ -9,7 +9,7 @@ import (
"github.bus.zalan.do/acid/postgres-operator/pkg/util" "github.bus.zalan.do/acid/postgres-operator/pkg/util"
) )
func (c *Cluster) clusterPods() ([]v1.Pod, error) { func (c *Cluster) listPods() ([]v1.Pod, error) {
ns := c.Metadata.Namespace ns := c.Metadata.Namespace
listOptions := v1.ListOptions{ listOptions := v1.ListOptions{
LabelSelector: c.labelsSet().String(), LabelSelector: c.labelsSet().String(),
@ -17,29 +17,12 @@ func (c *Cluster) clusterPods() ([]v1.Pod, error) {
pods, err := c.config.KubeClient.Pods(ns).List(listOptions) pods, err := c.config.KubeClient.Pods(ns).List(listOptions)
if err != nil { if err != nil {
return nil, fmt.Errorf("Can't get list of pods: %s", err) return nil, fmt.Errorf("Can't get list of Pods: %s", err)
} }
return pods.Items, nil return pods.Items, nil
} }
func (c *Cluster) deletePods() error {
pods, err := c.clusterPods()
if err != nil {
return err
}
for _, obj := range pods {
if err := c.deletePod(&obj); err != nil {
c.logger.Errorf("Can't delete pod: %s", err)
} else {
c.logger.Infof("Pod '%s' has been deleted", util.NameFromMeta(obj.ObjectMeta))
}
}
return nil
}
func (c *Cluster) listPersistentVolumeClaims() ([]v1.PersistentVolumeClaim, error) { func (c *Cluster) listPersistentVolumeClaims() ([]v1.PersistentVolumeClaim, error) {
ns := c.Metadata.Namespace ns := c.Metadata.Namespace
listOptions := v1.ListOptions{ listOptions := v1.ListOptions{
@ -53,6 +36,23 @@ func (c *Cluster) listPersistentVolumeClaims() ([]v1.PersistentVolumeClaim, erro
return pvcs.Items, nil return pvcs.Items, nil
} }
func (c *Cluster) deletePods() error {
pods, err := c.listPods()
if err != nil {
return err
}
for _, obj := range pods {
if err := c.deletePod(&obj); err != nil {
c.logger.Errorf("Can't delete Pod: %s", err)
} else {
c.logger.Infof("Pod '%s' has been deleted", util.NameFromMeta(obj.ObjectMeta))
}
}
return nil
}
func (c *Cluster) deletePersistenVolumeClaims() error { func (c *Cluster) deletePersistenVolumeClaims() error {
ns := c.Metadata.Namespace ns := c.Metadata.Namespace
pvcs, err := c.listPersistentVolumeClaims() pvcs, err := c.listPersistentVolumeClaims()
@ -116,7 +116,7 @@ func (c *Cluster) recreatePod(pod v1.Pod, spiloRole string) error {
}() }()
if err := c.config.KubeClient.Pods(pod.Namespace).Delete(pod.Name, deleteOptions); err != nil { if err := c.config.KubeClient.Pods(pod.Namespace).Delete(pod.Name, deleteOptions); err != nil {
return fmt.Errorf("Can't delete pod: %s", err) return fmt.Errorf("Can't delete Pod: %s", err)
} }
if err := c.waitForPodDeletion(ch); err != nil { if err := c.waitForPodDeletion(ch); err != nil {
@ -136,10 +136,10 @@ func (c *Cluster) podEventsDispatcher(stopCh <-chan struct{}) {
select { select {
case event := <-c.podEvents: case event := <-c.podEvents:
if subscriber, ok := c.podSubscribers[event.PodName]; ok { if subscriber, ok := c.podSubscribers[event.PodName]; ok {
c.logger.Debugf("New event for '%s' pod", event.PodName) c.logger.Debugf("New event for '%s' Pod", event.PodName)
go func() { subscriber <- event }() //TODO: is it a right way to do nonblocking send to the channel? go func() { subscriber <- event }() //TODO: is it a right way to do nonblocking send to the channel?
} else { } else {
c.logger.Debugf("Skipping event for an unwatched pod '%s'", event.PodName) c.logger.Debugf("Skipping event for an unwatched Pod '%s'", event.PodName)
} }
case <-stopCh: case <-stopCh:
return return
@ -156,15 +156,15 @@ func (c *Cluster) recreatePods() error {
} }
pods, err := c.config.KubeClient.Pods(namespace).List(listOptions) pods, err := c.config.KubeClient.Pods(namespace).List(listOptions)
if err != nil { if err != nil {
return fmt.Errorf("Can't get the list of the pods: %s", err) return fmt.Errorf("Can't get the list of Pods: %s", err)
} else { } else {
c.logger.Infof("There are %d pods in the cluster to recreate", len(pods.Items)) c.logger.Infof("There are %d Pods in the cluster to recreate", len(pods.Items))
} }
var masterPod v1.Pod var masterPod v1.Pod
for _, pod := range pods.Items { for _, pod := range pods.Items {
role, ok := pod.Labels["spilo-role"] role := util.PodSpiloRole(&pod)
if !ok { if role == "" {
continue continue
} }
@ -175,16 +175,19 @@ func (c *Cluster) recreatePods() error {
err = c.recreatePod(pod, "replica") err = c.recreatePod(pod, "replica")
if err != nil { if err != nil {
return fmt.Errorf("Can't recreate replica pod '%s': %s", util.NameFromMeta(pod.ObjectMeta), err) return fmt.Errorf("Can't recreate replica Pod '%s': %s", util.NameFromMeta(pod.ObjectMeta), err)
} }
} }
if masterPod.Name == "" {
c.logger.Warningln("No master Pod in the cluster")
}
//TODO: do manual failover //TODO: do manual failover
//TODO: specify master, leave new master empty //TODO: specify master, leave new master empty
c.logger.Infof("Recreating master pod '%s'", util.NameFromMeta(masterPod.ObjectMeta)) c.logger.Infof("Recreating master Pod '%s'", util.NameFromMeta(masterPod.ObjectMeta))
if err := c.recreatePod(masterPod, "replica"); err != nil { if err := c.recreatePod(masterPod, "replica"); err != nil {
return fmt.Errorf("Can't recreate master pod '%s': %s", util.NameFromMeta(masterPod.ObjectMeta), err) return fmt.Errorf("Can't recreate master Pod '%s': %s", util.NameFromMeta(masterPod.ObjectMeta), err)
} }
return nil return nil

View File

@ -18,7 +18,7 @@ var (
orphanDependents = false orphanDependents = false
) )
func getStatefulSet(clusterName spec.ClusterName, cSpec spec.PostgresSpec, etcdHost, dockerImage string) *v1beta1.StatefulSet { func genStatefulSet(clusterName spec.ClusterName, cSpec spec.PostgresSpec, etcdHost, dockerImage string) *v1beta1.StatefulSet {
volumeSize := cSpec.Volume.Size volumeSize := cSpec.Volume.Size
volumeStorageClass := cSpec.Volume.StorageClass volumeStorageClass := cSpec.Volume.StorageClass
resourceList := resources.ResourceList(cSpec.Resources) resourceList := resources.ResourceList(cSpec.Resources)
@ -36,7 +36,7 @@ func (c *Cluster) LoadResources() error {
services, err := c.config.KubeClient.Services(ns).List(listOptions) services, err := c.config.KubeClient.Services(ns).List(listOptions)
if err != nil { if err != nil {
return fmt.Errorf("Can't get list of services: %s", err) return fmt.Errorf("Can't get list of Services: %s", err)
} }
if len(services.Items) > 1 { if len(services.Items) > 1 {
return fmt.Errorf("Too many(%d) Services for a cluster", len(services.Items)) return fmt.Errorf("Too many(%d) Services for a cluster", len(services.Items))
@ -45,7 +45,7 @@ func (c *Cluster) LoadResources() error {
endpoints, err := c.config.KubeClient.Endpoints(ns).List(listOptions) endpoints, err := c.config.KubeClient.Endpoints(ns).List(listOptions)
if err != nil { if err != nil {
return fmt.Errorf("Can't get list of endpoints: %s", err) return fmt.Errorf("Can't get list of Endpoints: %s", err)
} }
if len(endpoints.Items) > 1 { if len(endpoints.Items) > 1 {
return fmt.Errorf("Too many(%d) Endpoints for a cluster", len(endpoints.Items)) return fmt.Errorf("Too many(%d) Endpoints for a cluster", len(endpoints.Items))
@ -54,7 +54,7 @@ func (c *Cluster) LoadResources() error {
secrets, err := c.config.KubeClient.Secrets(ns).List(listOptions) secrets, err := c.config.KubeClient.Secrets(ns).List(listOptions)
if err != nil { if err != nil {
return fmt.Errorf("Can't get list of secrets: %s", err) return fmt.Errorf("Can't get list of Secrets: %s", err)
} }
for i, secret := range secrets.Items { for i, secret := range secrets.Items {
if _, ok := c.Secrets[secret.UID]; ok { if _, ok := c.Secrets[secret.UID]; ok {
@ -66,7 +66,7 @@ func (c *Cluster) LoadResources() error {
statefulSets, err := c.config.KubeClient.StatefulSets(ns).List(listOptions) statefulSets, err := c.config.KubeClient.StatefulSets(ns).List(listOptions)
if err != nil { if err != nil {
return fmt.Errorf("Can't get list of stateful sets: %s", err) return fmt.Errorf("Can't get list of StatefulSets: %s", err)
} }
if len(statefulSets.Items) > 1 { if len(statefulSets.Items) > 1 {
return fmt.Errorf("Too many(%d) StatefulSets for a cluster", len(statefulSets.Items)) return fmt.Errorf("Too many(%d) StatefulSets for a cluster", len(statefulSets.Items))
@ -86,15 +86,24 @@ func (c *Cluster) ListResources() error {
c.logger.Infof("Endpoint: %s (uid: %s)", util.NameFromMeta(c.Endpoint.ObjectMeta), c.Endpoint.UID) c.logger.Infof("Endpoint: %s (uid: %s)", util.NameFromMeta(c.Endpoint.ObjectMeta), c.Endpoint.UID)
c.logger.Infof("Service: %s (uid: %s)", util.NameFromMeta(c.Service.ObjectMeta), c.Service.UID) c.logger.Infof("Service: %s (uid: %s)", util.NameFromMeta(c.Service.ObjectMeta), c.Service.UID)
pods, err := c.clusterPods() pods, err := c.listPods()
if err != nil { if err != nil {
return fmt.Errorf("Can't get pods: %s", err) return fmt.Errorf("Can't get the list of Pods: %s", err)
} }
for _, obj := range pods { for _, obj := range pods {
c.logger.Infof("Pod: %s (uid: %s)", util.NameFromMeta(obj.ObjectMeta), obj.UID) c.logger.Infof("Pod: %s (uid: %s)", util.NameFromMeta(obj.ObjectMeta), obj.UID)
} }
pvcs, err := c.listPersistentVolumeClaims()
if err != nil {
return fmt.Errorf("Can't get the list of PVCs: %s", err)
}
for _, obj := range pvcs {
c.logger.Infof("PVC: %s (uid: %s)", util.NameFromMeta(obj.ObjectMeta), obj.UID)
}
return nil return nil
} }
@ -102,7 +111,7 @@ func (c *Cluster) createStatefulSet() (*v1beta1.StatefulSet, error) {
if c.Statefulset != nil { if c.Statefulset != nil {
return nil, fmt.Errorf("StatefulSet already exists in the cluster") return nil, fmt.Errorf("StatefulSet already exists in the cluster")
} }
statefulSetSpec := getStatefulSet(c.ClusterName(), c.Spec, c.etcdHost, c.dockerImage) statefulSetSpec := genStatefulSet(c.ClusterName(), c.Spec, c.etcdHost, c.dockerImage)
statefulSet, err := c.config.KubeClient.StatefulSets(statefulSetSpec.Namespace).Create(statefulSetSpec) statefulSet, err := c.config.KubeClient.StatefulSets(statefulSetSpec.Namespace).Create(statefulSetSpec)
if k8sutil.ResourceAlreadyExists(err) { if k8sutil.ResourceAlreadyExists(err) {
return nil, fmt.Errorf("StatefulSet '%s' already exists", util.NameFromMeta(statefulSetSpec.ObjectMeta)) return nil, fmt.Errorf("StatefulSet '%s' already exists", util.NameFromMeta(statefulSetSpec.ObjectMeta))
@ -232,7 +241,7 @@ func (c *Cluster) applySecrets() error {
secrets, err := resources.UserSecrets(c.ClusterName(), c.pgUsers) secrets, err := resources.UserSecrets(c.ClusterName(), c.pgUsers)
if err != nil { if err != nil {
return fmt.Errorf("Can't get user secrets") return fmt.Errorf("Can't get user Secrets")
} }
for secretUsername, secretSpec := range secrets { for secretUsername, secretSpec := range secrets {
@ -240,7 +249,7 @@ func (c *Cluster) applySecrets() error {
if k8sutil.ResourceAlreadyExists(err) { if k8sutil.ResourceAlreadyExists(err) {
curSecrets, err := c.config.KubeClient.Secrets(secretSpec.Namespace).Get(secretSpec.Name) curSecrets, err := c.config.KubeClient.Secrets(secretSpec.Namespace).Get(secretSpec.Name)
if err != nil { if err != nil {
return fmt.Errorf("Can't get current secret: %s", err) return fmt.Errorf("Can't get current Secret: %s", err)
} }
pwdUser := c.pgUsers[secretUsername] pwdUser := c.pgUsers[secretUsername]
pwdUser.Password = string(curSecrets.Data["password"]) pwdUser.Password = string(curSecrets.Data["password"])
@ -249,10 +258,10 @@ func (c *Cluster) applySecrets() error {
continue continue
} else { } else {
if err != nil { if err != nil {
return fmt.Errorf("Can't create secret for user '%s': %s", secretUsername, err) return fmt.Errorf("Can't create Secret for user '%s': %s", secretUsername, err)
} }
c.Secrets[secret.UID] = secret c.Secrets[secret.UID] = secret
c.logger.Debugf("Created new secret, uid: %s", secret.UID) c.logger.Debugf("Created new Secret, uid: %s", secret.UID)
} }
} }

159
pkg/cluster/sync.go Normal file
View File

@ -0,0 +1,159 @@
package cluster
import (
"fmt"
"k8s.io/client-go/pkg/api/v1"
"github.bus.zalan.do/acid/postgres-operator/pkg/util"
"github.bus.zalan.do/acid/postgres-operator/pkg/util/resources"
)
func (c *Cluster) SyncCluster() {
if err := c.syncSecrets(); err != nil {
c.logger.Infof("Can't sync Secrets: %s", err)
}
if err := c.syncEndpoint(); err != nil {
c.logger.Errorf("Can't sync Endpoints: %s", err)
}
if err := c.syncService(); err != nil {
c.logger.Errorf("Can't sync Services: %s", err)
}
if err := c.syncStatefulSet(); err != nil {
c.logger.Errorf("Can't sync StatefulSets: %s", err)
}
if err := c.syncPods(); err != nil {
c.logger.Errorf("Can't sync Pods: %s", err)
}
}
func (c *Cluster) syncSecrets() error {
//TODO: mind the secrets of the deleted/new users
err := c.applySecrets()
if err != nil {
return err
}
return nil
}
func (c *Cluster) syncService() error {
cSpec := c.Spec
if c.Service == nil {
c.logger.Infof("Can't find the cluster's Service")
svc, err := c.createService()
if err != nil {
return fmt.Errorf("Can't create missing Service: %s", err)
}
c.logger.Infof("Created missing Service '%s'", util.NameFromMeta(svc.ObjectMeta))
return nil
}
desiredSvc := resources.Service(c.ClusterName(), c.Spec.TeamId, cSpec.AllowedSourceRanges)
if servicesEqual(c.Service, desiredSvc) {
return nil
}
c.logger.Infof("Service '%s' needs to be updated", util.NameFromMeta(desiredSvc.ObjectMeta))
if err := c.updateService(desiredSvc); err != nil {
return fmt.Errorf("Can't update Service to match desired state: %s", err)
}
c.logger.Infof("Service '%s' is in the desired state now", util.NameFromMeta(desiredSvc.ObjectMeta))
return nil
}
func (c *Cluster) syncEndpoint() error {
if c.Endpoint == nil {
c.logger.Infof("Can't find the cluster's Endpoint")
ep, err := c.createEndpoint()
if err != nil {
return fmt.Errorf("Can't create missing Endpoint: %s", err)
}
c.logger.Infof("Created missing Endpoint '%s'", util.NameFromMeta(ep.ObjectMeta))
return nil
}
return nil
}
func (c *Cluster) syncStatefulSet() error {
cSpec := c.Spec
if c.Statefulset == nil {
c.logger.Infof("Can't find the cluster's StatefulSet")
ss, err := c.createStatefulSet()
if err != nil {
return fmt.Errorf("Can't create missing StatefulSet: %s", err)
}
err = c.waitStatefulsetPodsReady()
if err != nil {
return fmt.Errorf("Cluster is not ready: %s", err)
}
c.logger.Infof("Created missing StatefulSet '%s'", util.NameFromMeta(ss.ObjectMeta))
return nil
}
desiredSS := genStatefulSet(c.ClusterName(), cSpec, c.etcdHost, c.dockerImage)
equalSS, rollUpdate := statefulsetsEqual(c.Statefulset, desiredSS)
if equalSS {
return nil
}
c.logger.Infof("StatefulSet '%s' is not in the desired state", util.NameFromMeta(c.Statefulset.ObjectMeta))
if err := c.updateStatefulSet(desiredSS); err != nil {
return fmt.Errorf("Can't update StatefulSet: %s", err)
}
if !rollUpdate {
c.logger.Debugln("No rolling update is needed")
return nil
}
c.logger.Debugln("Performing rolling update")
if err := c.recreatePods(); err != nil {
return fmt.Errorf("Can't recreate Pods: %s", err)
}
c.logger.Infof("Pods have been recreated")
return nil
}
func (c *Cluster) syncPods() error {
curSs := c.Statefulset
ls := c.labelsSet()
namespace := c.Metadata.Namespace
listOptions := v1.ListOptions{
LabelSelector: ls.String(),
}
pods, err := c.config.KubeClient.Pods(namespace).List(listOptions)
if err != nil {
return fmt.Errorf("Can't get list of Pods: %s", err)
}
if int32(len(pods.Items)) != *curSs.Spec.Replicas {
return fmt.Errorf("Number of existing Pods does not match number of replicas of the StatefulSet")
}
for _, pod := range pods.Items {
if podMatchesTemplate(&pod, curSs) {
c.logger.Infof("Pod '%s' matches StatefulSet pod template", util.NameFromMeta(pod.ObjectMeta))
continue
}
c.logger.Infof("Pod '%s' does not match StatefulSet pod template and needs to be deleted.", util.NameFromMeta(pod.ObjectMeta))
if util.PodSpiloRole(&pod) == "master" {
//TODO: do manual failover first
}
err = c.recreatePod(pod, "replica") // newly created pods are always "replica"s
c.logger.Infof("Pod '%s' has been successfully recreated", util.NameFromMeta(pod.ObjectMeta))
}
return nil
}

View File

@ -9,9 +9,11 @@ import (
etcdclient "github.com/coreos/etcd/client" etcdclient "github.com/coreos/etcd/client"
"k8s.io/client-go/pkg/api/v1" "k8s.io/client-go/pkg/api/v1"
"k8s.io/client-go/pkg/apis/apps/v1beta1"
"k8s.io/client-go/pkg/labels" "k8s.io/client-go/pkg/labels"
"github.bus.zalan.do/acid/postgres-operator/pkg/spec" "github.bus.zalan.do/acid/postgres-operator/pkg/spec"
"github.bus.zalan.do/acid/postgres-operator/pkg/util"
"github.bus.zalan.do/acid/postgres-operator/pkg/util/constants" "github.bus.zalan.do/acid/postgres-operator/pkg/util/constants"
"github.bus.zalan.do/acid/postgres-operator/pkg/util/retryutil" "github.bus.zalan.do/acid/postgres-operator/pkg/util/retryutil"
) )
@ -43,6 +45,83 @@ func normalizeUserFlags(userFlags []string) (flags []string, err error) {
return return
} }
func statefulsetsEqual(ss1, ss2 *v1beta1.StatefulSet) (equal bool, needsRollUpdate bool) {
equal = true
needsRollUpdate = false
//TODO: improve me
if *ss1.Spec.Replicas != *ss2.Spec.Replicas {
equal = false
}
if len(ss1.Spec.Template.Spec.Containers) != len(ss1.Spec.Template.Spec.Containers) {
equal = false
needsRollUpdate = true
return
}
if len(ss1.Spec.Template.Spec.Containers) == 0 {
return
}
container1 := ss1.Spec.Template.Spec.Containers[0]
container2 := ss2.Spec.Template.Spec.Containers[0]
if container1.Image != container2.Image {
equal = false
needsRollUpdate = true
return
}
if !reflect.DeepEqual(container1.Ports, container2.Ports) {
equal = false
needsRollUpdate = true
return
}
if !reflect.DeepEqual(container1.Resources, container2.Resources) {
equal = false
needsRollUpdate = true
return
}
if !reflect.DeepEqual(container1.Env, container2.Env) {
equal = false
needsRollUpdate = true
}
return
}
func servicesEqual(svc1, svc2 *v1.Service) bool {
//TODO: check of Ports
//TODO: improve me
if reflect.DeepEqual(svc1.Spec.LoadBalancerSourceRanges, svc2.Spec.LoadBalancerSourceRanges) {
return true
}
return false
}
func podMatchesTemplate(pod *v1.Pod, ss *v1beta1.StatefulSet) bool {
//TODO: improve me
if len(pod.Spec.Containers) != 1 {
return false
}
container := pod.Spec.Containers[0]
ssContainer := ss.Spec.Template.Spec.Containers[0]
if container.Image != ssContainer.Image {
return false
}
if !reflect.DeepEqual(container.Env, ssContainer.Env) {
return false
}
if !reflect.DeepEqual(container.Ports, ssContainer.Ports) {
return false
}
if !reflect.DeepEqual(container.Resources, ssContainer.Resources) {
return false
}
return true
}
func (c *Cluster) getTeamMembers() ([]string, error) { func (c *Cluster) getTeamMembers() ([]string, error) {
teamInfo, err := c.config.TeamsAPIClient.TeamInfo(c.Spec.TeamId) teamInfo, err := c.config.TeamsAPIClient.TeamInfo(c.Spec.TeamId)
if err != nil { if err != nil {
@ -56,10 +135,8 @@ func (c *Cluster) waitForPodLabel(podEvents chan spec.PodEvent, spiloRole string
for { for {
select { select {
case podEvent := <-podEvents: case podEvent := <-podEvents:
podLabels := podEvent.CurPod.Labels role := util.PodSpiloRole(podEvent.CurPod)
c.logger.Debugf("Pod has following labels: %+v", podLabels) if role == spiloRole {
val, ok := podLabels["spilo-role"]
if ok && val == spiloRole {
return nil return nil
} }
case <-time.After(constants.PodLabelWaitTimeout): case <-time.After(constants.PodLabelWaitTimeout):
@ -82,7 +159,7 @@ func (c *Cluster) waitForPodDeletion(podEvents chan spec.PodEvent) error {
} }
func (c *Cluster) waitStatefulsetReady() error { func (c *Cluster) waitStatefulsetReady() error {
return retryutil.Retry(constants.ResourceCheckInterval, int(constants.ResourceCheckTimeout/constants.ResourceCheckInterval), return retryutil.Retry(constants.ResourceCheckInterval, constants.ResourceCheckTimeout,
func() (bool, error) { func() (bool, error) {
listOptions := v1.ListOptions{ listOptions := v1.ListOptions{
LabelSelector: c.labelsSet().String(), LabelSelector: c.labelsSet().String(),
@ -119,8 +196,7 @@ func (c *Cluster) waitPodLabelsReady() error {
} }
podsNumber := len(pods.Items) podsNumber := len(pods.Items)
return retryutil.Retry( return retryutil.Retry(constants.ResourceCheckInterval, constants.ResourceCheckTimeout,
constants.ResourceCheckInterval, int(constants.ResourceCheckTimeout/constants.ResourceCheckInterval),
func() (bool, error) { func() (bool, error) {
masterPods, err := c.config.KubeClient.Pods(namespace).List(masterListOption) masterPods, err := c.config.KubeClient.Pods(namespace).List(masterListOption)
if err != nil { if err != nil {
@ -133,6 +209,9 @@ func (c *Cluster) waitPodLabelsReady() error {
if len(masterPods.Items) > 1 { if len(masterPods.Items) > 1 {
return false, fmt.Errorf("Too many masters") return false, fmt.Errorf("Too many masters")
} }
if len(replicaPods.Items) == podsNumber {
return false, fmt.Errorf("Cluster has no master")
}
return len(masterPods.Items)+len(replicaPods.Items) == podsNumber, nil return len(masterPods.Items)+len(replicaPods.Items) == podsNumber, nil
}) })
@ -185,12 +264,3 @@ func (c *Cluster) deleteEtcdKey() error {
return nil return nil
} }
func servicesEqual(svc1, svc2 *v1.Service) bool {
//TODO: improve me
if reflect.DeepEqual(svc1.Spec.LoadBalancerSourceRanges, svc2.Spec.LoadBalancerSourceRanges) {
return true
}
return false
}

View File

@ -117,15 +117,15 @@ func (c *Controller) podDelete(obj interface{}) {
} }
func (c *Controller) podEventsDispatcher(stopCh <-chan struct{}) { func (c *Controller) podEventsDispatcher(stopCh <-chan struct{}) {
c.logger.Infof("Watching all pod events") c.logger.Infof("Watching all Pod events")
for { for {
select { select {
case event := <-c.podCh: case event := <-c.podCh:
if subscriber, ok := c.clusters[event.ClusterName]; ok { if subscriber, ok := c.clusters[event.ClusterName]; ok {
c.logger.Debugf("Sending %s event of pod '%s' to the '%s' cluster channel", event.EventType, event.PodName, event.ClusterName) c.logger.Debugf("Sending %s event of Pod '%s' to the '%s' cluster channel", event.EventType, event.PodName, event.ClusterName)
go subscriber.ReceivePodEvent(event) go subscriber.ReceivePodEvent(event)
} else { } else {
c.logger.Debugf("Skipping pods unrelated to clusters: %s", event.PodName) c.logger.Debugf("Skipping Pods unrelated to clusters: %s", event.PodName)
} }
case <-stopCh: case <-stopCh:
return return

View File

@ -53,7 +53,9 @@ func (c *Controller) clusterListFunc(options api.ListOptions) (runtime.Object, e
c.clusters[clusterName] = cl c.clusters[clusterName] = cl
cl.LoadResources() cl.LoadResources()
go cl.Run(stopCh) go cl.Run(stopCh)
cl.ListResources() cl.ListResources()
cl.SyncCluster()
} }
if len(c.clusters) > 0 { if len(c.clusters) > 0 {
c.logger.Infof("There are %d clusters currently running", len(c.clusters)) c.logger.Infof("There are %d clusters currently running", len(c.clusters))

View File

@ -24,7 +24,7 @@ func (c *Controller) getOAuthToken() (string, error) {
credentialsSecret, err := c.config.KubeClient.Secrets(api.NamespaceDefault).Get(constants.OAuthTokenSecretName) credentialsSecret, err := c.config.KubeClient.Secrets(api.NamespaceDefault).Get(constants.OAuthTokenSecretName)
if err != nil { if err != nil {
return "", fmt.Errorf("Can't get credentials secret: %s", err) return "", fmt.Errorf("Can't get credentials Secret: %s", err)
} }
data := credentialsSecret.Data data := credentialsSecret.Data

View File

@ -29,12 +29,13 @@ const (
PamConfiguration = "https://info.example.com/oauth2/tokeninfo?access_token= uid realm=/employees" PamConfiguration = "https://info.example.com/oauth2/tokeninfo?access_token= uid realm=/employees"
PasswordLength = 64 PasswordLength = 64
TeamsAPIUrl = "https://teams.example.com/api/" TeamsAPIUrl = "https://teams.example.com/api/"
UserSecretTemplate = "%s.%s.credentials.%s.%s" UserSecretTemplate = "%s.%s.credentials.%s.%s" // Username, ClusterName, TPRName, TPRVendor
OAuthTokenSecretName = "postgresql-operator" OAuthTokenSecretName = "postgresql-operator"
ServiceAccountName = "operator" ServiceAccountName = "operator"
DataVolumeName = "pgdata" DataVolumeName = "pgdata"
ZalandoDnsNameAnnotation = "zalando.org/dnsname" ZalandoDnsNameAnnotation = "zalando.org/dnsname"
// TODO: move DbHostedZone to operator configuration // TODO: move DbHostedZone to operator configuration
DbHostedZone = "db.example.com" DbHostedZone = "db.example.com"
) )

View File

@ -15,7 +15,7 @@ func listClustersURI(ns string) string {
} }
func WaitTPRReady(restclient rest.Interface, interval, timeout time.Duration, ns string) error { func WaitTPRReady(restclient rest.Interface, interval, timeout time.Duration, ns string) error {
return retryutil.Retry(interval, int(timeout/interval), func() (bool, error) { return retryutil.Retry(interval, timeout, func() (bool, error) {
_, err := restclient.Get().RequestURI(listClustersURI(ns)).DoRaw() _, err := restclient.Get().RequestURI(listClustersURI(ns)).DoRaw()
if err != nil { if err != nil {
if ResourceNotFound(err) { // not set up yet. wait more. if ResourceNotFound(err) { // not set up yet. wait more.

View File

@ -5,24 +5,14 @@ import (
"time" "time"
) )
type RetryError struct {
n int
}
func (e *RetryError) Error() string {
return fmt.Sprintf("still failing after %d retries", e.n)
}
type ConditionFunc func() (bool, error) type ConditionFunc func() (bool, error)
// Retry retries f every interval until after maxRetries. func Retry(interval time.Duration, timeout time.Duration, f ConditionFunc) error {
// The interval won't be affected by how long f takes. //TODO: make the retry exponential
// For example, if interval is 3s, f takes 1s, another f will be called 2s later. if timeout < interval {
// However, if f takes longer than interval, it will be delayed. return fmt.Errorf("timout(%s) should be greater than interval(%s)", timeout, interval)
func Retry(interval time.Duration, maxRetries int, f ConditionFunc) error {
if maxRetries <= 0 {
return fmt.Errorf("maxRetries (%d) should be > 0", maxRetries)
} }
maxRetries := int(timeout / interval)
tick := time.NewTicker(interval) tick := time.NewTicker(interval)
defer tick.Stop() defer tick.Stop()
@ -39,5 +29,5 @@ func Retry(interval time.Duration, maxRetries int, f ConditionFunc) error {
} }
<-tick.C <-tick.C
} }
return &RetryError{maxRetries} return fmt.Errorf("Still failing after %d retries", maxRetries)
} }

View File

@ -45,6 +45,10 @@ func PodClusterName(pod *v1.Pod) spec.ClusterName {
return spec.ClusterName{} return spec.ClusterName{}
} }
func PodSpiloRole(pod *v1.Pod) string {
return pod.Labels["spilo-role"]
}
func ClusterDNSName(clusterName, teamName, hostedZone string) string { func ClusterDNSName(clusterName, teamName, hostedZone string) string {
return fmt.Sprintf("%s.%s.%s", clusterName, teamName, hostedZone) return fmt.Sprintf("%s.%s.%s", clusterName, teamName, hostedZone)
} }