create CDC event stream CRD (#1570)
* provide event stream API * check manifest settings for logical decoding before creating streams * operator updates Postgres config and creates replication user * name FES like the Postgres cluster * add delete case and fix updating streams + update unit test * check if fes CRD exists before syncing * existing slot must use the same plugin * make id and payload columns configurable * sync streams only when they are defined in manifest * introduce applicationId for separate stream CRDs * add FES to RBAC in chart * disable streams in chart * switch to pgoutput plugin and let operator create publications * reflect code review and additional refactoring Co-authored-by: Paŭlo Ebermann <paul.ebermann@zalando.de>
This commit is contained in:
		
							parent
							
								
									8b404fd049
								
							
						
					
					
						commit
						d8a159ef1a
					
				|  | @ -474,6 +474,39 @@ spec: | ||||||
|                     type: string |                     type: string | ||||||
|                   gs_wal_path: |                   gs_wal_path: | ||||||
|                     type: string |                     type: string | ||||||
|  |               streams: | ||||||
|  |                 type: array | ||||||
|  |                 nullable: true | ||||||
|  |                 items: | ||||||
|  |                   type: object | ||||||
|  |                   required: | ||||||
|  |                     - applicationId | ||||||
|  |                     - database | ||||||
|  |                     - tables | ||||||
|  |                   properties: | ||||||
|  |                     applicationId: | ||||||
|  |                       type: string | ||||||
|  |                     batchSize: | ||||||
|  |                       type: integer | ||||||
|  |                     database: | ||||||
|  |                       type: string | ||||||
|  |                     filter: | ||||||
|  |                       type: object | ||||||
|  |                       additionalProperties: | ||||||
|  |                         type: string | ||||||
|  |                     tables: | ||||||
|  |                       type: object | ||||||
|  |                       additionalProperties: | ||||||
|  |                         type: object | ||||||
|  |                         required: | ||||||
|  |                           - eventType | ||||||
|  |                         properties: | ||||||
|  |                           eventType: | ||||||
|  |                             type: string | ||||||
|  |                           idColumn: | ||||||
|  |                             type: string | ||||||
|  |                           payloadColumn: | ||||||
|  |                             type: string | ||||||
|               teamId: |               teamId: | ||||||
|                 type: string |                 type: string | ||||||
|               tls: |               tls: | ||||||
|  |  | ||||||
|  | @ -34,6 +34,22 @@ rules: | ||||||
|   - get |   - get | ||||||
|   - list |   - list | ||||||
|   - watch |   - watch | ||||||
|  | # all verbs allowed for event streams | ||||||
|  | {{- if .Values.enableStreams }} | ||||||
|  | - apiGroups: | ||||||
|  |   - zalando.org | ||||||
|  |   resources: | ||||||
|  |   - fabriceventstreams | ||||||
|  |   verbs: | ||||||
|  |   - create | ||||||
|  |   - delete | ||||||
|  |   - deletecollection | ||||||
|  |   - get | ||||||
|  |   - list | ||||||
|  |   - patch | ||||||
|  |   - update | ||||||
|  |   - watch | ||||||
|  | {{- end }} | ||||||
| # to create or get/update CRDs when starting up | # to create or get/update CRDs when starting up | ||||||
| - apiGroups: | - apiGroups: | ||||||
|   - apiextensions.k8s.io |   - apiextensions.k8s.io | ||||||
|  |  | ||||||
|  | @ -368,6 +368,9 @@ configConnectionPooler: | ||||||
|   connection_pooler_default_cpu_limit: "1" |   connection_pooler_default_cpu_limit: "1" | ||||||
|   connection_pooler_default_memory_limit: 100Mi |   connection_pooler_default_memory_limit: 100Mi | ||||||
| 
 | 
 | ||||||
|  | # Zalando's internal CDC stream feature | ||||||
|  | enableStreams: false | ||||||
|  | 
 | ||||||
| rbac: | rbac: | ||||||
|   # Specifies whether RBAC resources should be created |   # Specifies whether RBAC resources should be created | ||||||
|   create: true |   create: true | ||||||
|  |  | ||||||
|  | @ -536,3 +536,54 @@ Those parameters are grouped under the `tls` top-level key. | ||||||
|   relative to the "/tls/", which is mount path of the tls secret. |   relative to the "/tls/", which is mount path of the tls secret. | ||||||
|   If `caSecretName` is defined, the ca.crt path is relative to "/tlsca/", |   If `caSecretName` is defined, the ca.crt path is relative to "/tlsca/", | ||||||
|   otherwise to the same "/tls/". |   otherwise to the same "/tls/". | ||||||
|  | 
 | ||||||
|  | ## Change data capture streams | ||||||
|  | 
 | ||||||
|  | This sections enables change data capture (CDC) streams via Postgres'  | ||||||
|  | [logical decoding](https://www.postgresql.org/docs/14/logicaldecoding.html) | ||||||
|  | feature and `pgoutput` plugin. While the Postgres operator takes responsibility | ||||||
|  | for providing the setup to publish change events, it relies on external tools | ||||||
|  | to consume them. At Zalando, we are using a workflow based on | ||||||
|  | [Debezium Connector](https://debezium.io/documentation/reference/stable/connectors/postgresql.html) | ||||||
|  | which can feed streams into Zalando’s distributed event broker [Nakadi](https://nakadi.io/) | ||||||
|  | among others. | ||||||
|  | 
 | ||||||
|  | The Postgres Operator creates custom resources for Zalando's internal CDC | ||||||
|  | operator which will be used to set up the consumer part. Each stream object | ||||||
|  | can have the following properties: | ||||||
|  | 
 | ||||||
|  | * **applicationId** | ||||||
|  |   The application name to which the database and CDC belongs to. For each | ||||||
|  |   set of streams with a distinct `applicationId` a separate stream CR as well | ||||||
|  |   as a separate logical replication slot will be created. This means there can | ||||||
|  |   be different streams in the same database and streams with the same | ||||||
|  |   `applicationId` are bundled in one stream CR. The stream CR will be called | ||||||
|  |   like the Postgres cluster plus "-<applicationId>" suffix. Required. | ||||||
|  | 
 | ||||||
|  | * **database** | ||||||
|  |   Name of the database from where events will be published via Postgres' | ||||||
|  |   logical decoding feature. The operator will take care of updating the | ||||||
|  |   database configuration (setting `wal_level: logical`, creating logical | ||||||
|  |   replication slots, using output plugin `pgoutput` and creating a dedicated | ||||||
|  |   replication user). Required. | ||||||
|  | 
 | ||||||
|  | * **tables** | ||||||
|  |   Defines a map of table names and their properties (`eventType`, `idColumn` | ||||||
|  |   and `payloadColumn`). The CDC operator is following the [outbox pattern](https://debezium.io/blog/2019/02/19/reliable-microservices-data-exchange-with-the-outbox-pattern/). | ||||||
|  |   The application is responsible for putting events into a (JSON/B or VARCHAR) | ||||||
|  |   payload column of the outbox table in the structure of the specified target | ||||||
|  |   event type. The operator will create a [PUBLICATION](https://www.postgresql.org/docs/14/logical-replication-publication.html) | ||||||
|  |   in Postgres for all tables specified for one `database` and `applicationId`. | ||||||
|  |   The CDC operator will consume from it shortly after transactions are | ||||||
|  |   committed to the outbox table. The `idColumn` will be used in telemetry for | ||||||
|  |   the CDC operator. The names for `idColumn` and `payloadColumn` can be | ||||||
|  |   configured. Defaults are `id` and `payload`. The target `eventType` has to | ||||||
|  |   be defined. Required. | ||||||
|  | 
 | ||||||
|  | * **filter** | ||||||
|  |   Streamed events can be filtered by a jsonpath expression for each table. | ||||||
|  |   Optional. | ||||||
|  | 
 | ||||||
|  | * **batchSize** | ||||||
|  |   Defines the size of batches in which events are consumed. Optional. | ||||||
|  |   Defaults to 1. | ||||||
|  |  | ||||||
|  | @ -17,7 +17,7 @@ trap "cleanup" EXIT SIGINT | ||||||
| 
 | 
 | ||||||
| bash "${CODEGEN_PKG}/generate-groups.sh" all \ | bash "${CODEGEN_PKG}/generate-groups.sh" all \ | ||||||
|   "${OPERATOR_PACKAGE_ROOT}/pkg/generated" "${OPERATOR_PACKAGE_ROOT}/pkg/apis" \ |   "${OPERATOR_PACKAGE_ROOT}/pkg/generated" "${OPERATOR_PACKAGE_ROOT}/pkg/apis" \ | ||||||
|   "acid.zalan.do:v1" \ |   "acid.zalan.do:v1 zalando.org:v1" \ | ||||||
|   --go-header-file "${SCRIPT_ROOT}"/hack/custom-boilerplate.go.txt |   --go-header-file "${SCRIPT_ROOT}"/hack/custom-boilerplate.go.txt | ||||||
| 
 | 
 | ||||||
| cp -r "${OPERATOR_PACKAGE_ROOT}"/pkg/* "${TARGET_CODE_DIR}" | cp -r "${OPERATOR_PACKAGE_ROOT}"/pkg/* "${TARGET_CODE_DIR}" | ||||||
|  |  | ||||||
|  | @ -3,6 +3,7 @@ kind: postgresql | ||||||
| metadata: | metadata: | ||||||
|   name: acid-test-cluster |   name: acid-test-cluster | ||||||
| #  labels: | #  labels: | ||||||
|  | #    application: test-app | ||||||
| #    environment: demo | #    environment: demo | ||||||
| #  annotations: | #  annotations: | ||||||
| #    "acid.zalan.do/controller": "second-operator" | #    "acid.zalan.do/controller": "second-operator" | ||||||
|  | @ -203,3 +204,21 @@ spec: | ||||||
| #              operator: In | #              operator: In | ||||||
| #              values: | #              values: | ||||||
| #                - enabled | #                - enabled | ||||||
|  | 
 | ||||||
|  | # Enables change data capture streams for defined database tables | ||||||
|  | #  streams: | ||||||
|  | #  - applicationId: test-app | ||||||
|  | #    database: foo | ||||||
|  | #    tables: | ||||||
|  | #      data.state_pending_outbox: | ||||||
|  | #        eventType: test-app.status-pending | ||||||
|  | #      data.state_approved_outbox: | ||||||
|  | #        eventType: test-app.status-approved | ||||||
|  | #      data.orders_outbox: | ||||||
|  | #        eventType: test-app.order-completed | ||||||
|  | #        idColumn: o_id | ||||||
|  | #        payloadColumn: o_payload | ||||||
|  | #    # Optional. Filter ignores events before a certain txnId and lsn. Can be used to skip bad events | ||||||
|  | #    filter: | ||||||
|  | #      data.orders_outbox: "[?(@.source.txId > 500 && @.source.lsn > 123456)]" | ||||||
|  | #    batchSize: 1000 | ||||||
|  |  | ||||||
|  | @ -35,6 +35,20 @@ rules: | ||||||
|   - get |   - get | ||||||
|   - list |   - list | ||||||
|   - watch |   - watch | ||||||
|  | # all verbs allowed for event streams (Zalando-internal feature) | ||||||
|  | # - apiGroups: | ||||||
|  | #   - zalando.org | ||||||
|  | #   resources: | ||||||
|  | #   - fabriceventstreams | ||||||
|  | #   verbs: | ||||||
|  | #   - create | ||||||
|  | #   - delete | ||||||
|  | #   - deletecollection | ||||||
|  | #   - get | ||||||
|  | #   - list | ||||||
|  | #   - patch | ||||||
|  | #   - update | ||||||
|  | #   - watch | ||||||
| # to create or get/update CRDs when starting up | # to create or get/update CRDs when starting up | ||||||
| - apiGroups: | - apiGroups: | ||||||
|   - apiextensions.k8s.io |   - apiextensions.k8s.io | ||||||
|  |  | ||||||
|  | @ -472,6 +472,39 @@ spec: | ||||||
|                     type: string |                     type: string | ||||||
|                   gs_wal_path: |                   gs_wal_path: | ||||||
|                     type: string |                     type: string | ||||||
|  |               streams: | ||||||
|  |                 type: array | ||||||
|  |                 nullable: true | ||||||
|  |                 items: | ||||||
|  |                   type: object | ||||||
|  |                   required: | ||||||
|  |                     - applicationId | ||||||
|  |                     - database | ||||||
|  |                     - tables | ||||||
|  |                   properties: | ||||||
|  |                     applicationId: | ||||||
|  |                       type: string | ||||||
|  |                     batchSize: | ||||||
|  |                       type: integer | ||||||
|  |                     database: | ||||||
|  |                       type: string | ||||||
|  |                     filter: | ||||||
|  |                       type: object | ||||||
|  |                       additionalProperties: | ||||||
|  |                         type: string | ||||||
|  |                     tables: | ||||||
|  |                       type: object | ||||||
|  |                       additionalProperties: | ||||||
|  |                         type: object | ||||||
|  |                         required: | ||||||
|  |                           - eventType | ||||||
|  |                         properties: | ||||||
|  |                           eventType: | ||||||
|  |                             type: string | ||||||
|  |                           idColumn: | ||||||
|  |                             type: string | ||||||
|  |                           payloadColumn: | ||||||
|  |                             type: string | ||||||
|               teamId: |               teamId: | ||||||
|                 type: string |                 type: string | ||||||
|               tls: |               tls: | ||||||
|  |  | ||||||
|  | @ -716,6 +716,54 @@ var PostgresCRDResourceValidation = apiextv1.CustomResourceValidation{ | ||||||
| 							}, | 							}, | ||||||
| 						}, | 						}, | ||||||
| 					}, | 					}, | ||||||
|  | 					"streams": { | ||||||
|  | 						Type: "array", | ||||||
|  | 						Items: &apiextv1.JSONSchemaPropsOrArray{ | ||||||
|  | 							Schema: &apiextv1.JSONSchemaProps{ | ||||||
|  | 								Type:     "object", | ||||||
|  | 								Required: []string{"applicationId", "database", "tables"}, | ||||||
|  | 								Properties: map[string]apiextv1.JSONSchemaProps{ | ||||||
|  | 									"applicationId": { | ||||||
|  | 										Type: "string", | ||||||
|  | 									}, | ||||||
|  | 									"batchSize": { | ||||||
|  | 										Type: "integer", | ||||||
|  | 									}, | ||||||
|  | 									"database": { | ||||||
|  | 										Type: "string", | ||||||
|  | 									}, | ||||||
|  | 									"filter": { | ||||||
|  | 										Type: "object", | ||||||
|  | 										AdditionalProperties: &apiextv1.JSONSchemaPropsOrBool{ | ||||||
|  | 											Schema: &apiextv1.JSONSchemaProps{ | ||||||
|  | 												Type: "string", | ||||||
|  | 											}, | ||||||
|  | 										}, | ||||||
|  | 									}, | ||||||
|  | 									"tables": { | ||||||
|  | 										Type: "object", | ||||||
|  | 										AdditionalProperties: &apiextv1.JSONSchemaPropsOrBool{ | ||||||
|  | 											Schema: &apiextv1.JSONSchemaProps{ | ||||||
|  | 												Type:     "object", | ||||||
|  | 												Required: []string{"eventType"}, | ||||||
|  | 												Properties: map[string]apiextv1.JSONSchemaProps{ | ||||||
|  | 													"eventType": { | ||||||
|  | 														Type: "string", | ||||||
|  | 													}, | ||||||
|  | 													"idColumn": { | ||||||
|  | 														Type: "string", | ||||||
|  | 													}, | ||||||
|  | 													"payloadColumn": { | ||||||
|  | 														Type: "string", | ||||||
|  | 													}, | ||||||
|  | 												}, | ||||||
|  | 											}, | ||||||
|  | 										}, | ||||||
|  | 									}, | ||||||
|  | 								}, | ||||||
|  | 							}, | ||||||
|  | 						}, | ||||||
|  | 					}, | ||||||
| 					"teamId": { | 					"teamId": { | ||||||
| 						Type: "string", | 						Type: "string", | ||||||
| 					}, | 					}, | ||||||
|  |  | ||||||
|  | @ -77,6 +77,7 @@ type PostgresSpec struct { | ||||||
| 	ServiceAnnotations    map[string]string           `json:"serviceAnnotations,omitempty"` | 	ServiceAnnotations    map[string]string           `json:"serviceAnnotations,omitempty"` | ||||||
| 	TLS                   *TLSDescription             `json:"tls,omitempty"` | 	TLS                   *TLSDescription             `json:"tls,omitempty"` | ||||||
| 	AdditionalVolumes     []AdditionalVolume          `json:"additionalVolumes,omitempty"` | 	AdditionalVolumes     []AdditionalVolume          `json:"additionalVolumes,omitempty"` | ||||||
|  | 	Streams               []Stream                    `json:"streams,omitempty"` | ||||||
| 
 | 
 | ||||||
| 	// deprecated json tags
 | 	// deprecated json tags
 | ||||||
| 	InitContainersOld       []v1.Container `json:"init_containers,omitempty"` | 	InitContainersOld       []v1.Container `json:"init_containers,omitempty"` | ||||||
|  | @ -231,3 +232,17 @@ type ConnectionPooler struct { | ||||||
| 
 | 
 | ||||||
| 	Resources `json:"resources,omitempty"` | 	Resources `json:"resources,omitempty"` | ||||||
| } | } | ||||||
|  | 
 | ||||||
|  | type Stream struct { | ||||||
|  | 	ApplicationId string                 `json:"applicationId"` | ||||||
|  | 	Database      string                 `json:"database"` | ||||||
|  | 	Tables        map[string]StreamTable `json:"tables"` | ||||||
|  | 	Filter        map[string]string      `json:"filter,omitempty"` | ||||||
|  | 	BatchSize     uint32                 `json:"batchSize,omitempty"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | type StreamTable struct { | ||||||
|  | 	EventType     string `json:"eventType"` | ||||||
|  | 	IdColumn      string `json:"idColumn,omitempty" defaults:"id"` | ||||||
|  | 	PayloadColumn string `json:"payloadColumn,omitempty" defaults:"payload"` | ||||||
|  | } | ||||||
|  |  | ||||||
|  | @ -749,6 +749,13 @@ func (in *PostgresSpec) DeepCopyInto(out *PostgresSpec) { | ||||||
| 			(*in)[i].DeepCopyInto(&(*out)[i]) | 			(*in)[i].DeepCopyInto(&(*out)[i]) | ||||||
| 		} | 		} | ||||||
| 	} | 	} | ||||||
|  | 	if in.Streams != nil { | ||||||
|  | 		in, out := &in.Streams, &out.Streams | ||||||
|  | 		*out = make([]Stream, len(*in)) | ||||||
|  | 		for i := range *in { | ||||||
|  | 			(*in)[i].DeepCopyInto(&(*out)[i]) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
| 	if in.InitContainersOld != nil { | 	if in.InitContainersOld != nil { | ||||||
| 		in, out := &in.InitContainersOld, &out.InitContainersOld | 		in, out := &in.InitContainersOld, &out.InitContainersOld | ||||||
| 		*out = make([]corev1.Container, len(*in)) | 		*out = make([]corev1.Container, len(*in)) | ||||||
|  | @ -1152,6 +1159,52 @@ func (in *StandbyDescription) DeepCopy() *StandbyDescription { | ||||||
| 	return out | 	return out | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *Stream) DeepCopyInto(out *Stream) { | ||||||
|  | 	*out = *in | ||||||
|  | 	if in.Tables != nil { | ||||||
|  | 		in, out := &in.Tables, &out.Tables | ||||||
|  | 		*out = make(map[string]StreamTable, len(*in)) | ||||||
|  | 		for key, val := range *in { | ||||||
|  | 			(*out)[key] = val | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 	if in.Filter != nil { | ||||||
|  | 		in, out := &in.Filter, &out.Filter | ||||||
|  | 		*out = make(map[string]string, len(*in)) | ||||||
|  | 		for key, val := range *in { | ||||||
|  | 			(*out)[key] = val | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Stream.
 | ||||||
|  | func (in *Stream) DeepCopy() *Stream { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(Stream) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *StreamTable) DeepCopyInto(out *StreamTable) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new StreamTable.
 | ||||||
|  | func (in *StreamTable) DeepCopy() *StreamTable { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(StreamTable) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
| // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
| func (in *TLSDescription) DeepCopyInto(out *TLSDescription) { | func (in *TLSDescription) DeepCopyInto(out *TLSDescription) { | ||||||
| 	*out = *in | 	*out = *in | ||||||
|  |  | ||||||
|  | @ -0,0 +1,6 @@ | ||||||
|  | package zalando | ||||||
|  | 
 | ||||||
|  | const ( | ||||||
|  | 	// GroupName is the group name for the operator CRDs
 | ||||||
|  | 	GroupName = "zalando.org" | ||||||
|  | ) | ||||||
|  | @ -0,0 +1,84 @@ | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // +genclient
 | ||||||
|  | // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
 | ||||||
|  | 
 | ||||||
|  | // FabricEventStream defines FabricEventStream Custom Resource Definition Object.
 | ||||||
|  | type FabricEventStream struct { | ||||||
|  | 	metav1.TypeMeta   `json:",inline"` | ||||||
|  | 	metav1.ObjectMeta `json:"metadata,omitempty"` | ||||||
|  | 
 | ||||||
|  | 	Spec FabricEventStreamSpec `json:"spec"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamSpec defines the specification for the FabricEventStream TPR.
 | ||||||
|  | type FabricEventStreamSpec struct { | ||||||
|  | 	ApplicationId string        `json:"applicationId"` | ||||||
|  | 	EventStreams  []EventStream `json:"eventStreams"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object
 | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamList defines a list of FabricEventStreams .
 | ||||||
|  | type FabricEventStreamList struct { | ||||||
|  | 	metav1.TypeMeta `json:",inline"` | ||||||
|  | 	metav1.ListMeta `json:"metadata"` | ||||||
|  | 
 | ||||||
|  | 	Items []FabricEventStream `json:"items"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // EventStream defines the source, flow and sink of the event stream
 | ||||||
|  | type EventStream struct { | ||||||
|  | 	EventStreamFlow   EventStreamFlow   `json:"flow"` | ||||||
|  | 	EventStreamSink   EventStreamSink   `json:"sink"` | ||||||
|  | 	EventStreamSource EventStreamSource `json:"source"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // EventStreamFlow defines the flow characteristics of the event stream
 | ||||||
|  | type EventStreamFlow struct { | ||||||
|  | 	Type          string `json:"type"` | ||||||
|  | 	PayloadColumn string `json:"payloadColumn,omitempty" defaults:"payload"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // EventStreamSink defines the target of the event stream
 | ||||||
|  | type EventStreamSink struct { | ||||||
|  | 	Type         string `json:"type"` | ||||||
|  | 	EventType    string `json:"eventType,omitempty"` | ||||||
|  | 	MaxBatchSize uint32 `json:"maxBatchSize,omitempty"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // EventStreamSource defines the source of the event stream and connection for FES operator
 | ||||||
|  | type EventStreamSource struct { | ||||||
|  | 	Type             string           `json:"type"` | ||||||
|  | 	Schema           string           `json:"schema,omitempty" defaults:"public"` | ||||||
|  | 	EventStreamTable EventStreamTable `json:"table"` | ||||||
|  | 	Filter           string           `json:"filter,omitempty"` | ||||||
|  | 	Connection       Connection       `json:"jdbcConnection"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // EventStreamTable defines the name and ID column to be used for streaming
 | ||||||
|  | type EventStreamTable struct { | ||||||
|  | 	Name     string `json:"name"` | ||||||
|  | 	IDColumn string `json:"idColumn,omitempty" defaults:"id"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Connection to be used for allowing the FES operator to connect to a database
 | ||||||
|  | type Connection struct { | ||||||
|  | 	Url             string `json:"jdbcUrl"` | ||||||
|  | 	SlotName        string `json:"slotName"` | ||||||
|  | 	PluginType      string `json:"pluginType,omitempty" defaults:"wal2json"` | ||||||
|  | 	PublicationName string `json:"publicationName,omitempty"` | ||||||
|  | 	DBAuth          DBAuth `json:"databaseAuthentication"` | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DBAuth specifies the credentials to be used for connecting with the database
 | ||||||
|  | type DBAuth struct { | ||||||
|  | 	Type        string `json:"type"` | ||||||
|  | 	Name        string `json:"name,omitempty"` | ||||||
|  | 	UserKey     string `json:"userKey,omitempty"` | ||||||
|  | 	PasswordKey string `json:"passwordKey,omitempty"` | ||||||
|  | } | ||||||
|  | @ -0,0 +1,44 @@ | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/apis/zalando.org" | ||||||
|  | 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
|  | 	"k8s.io/apimachinery/pkg/runtime" | ||||||
|  | 	"k8s.io/apimachinery/pkg/runtime/schema" | ||||||
|  | 	"k8s.io/client-go/kubernetes/scheme" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // APIVersion of the `fabriceventstream` CRD
 | ||||||
|  | const ( | ||||||
|  | 	APIVersion = "v1" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | var ( | ||||||
|  | 	schemeBuilder = runtime.NewSchemeBuilder(addKnownTypes) | ||||||
|  | 	AddToScheme   = schemeBuilder.AddToScheme | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | func init() { | ||||||
|  | 	err := AddToScheme(scheme.Scheme) | ||||||
|  | 	if err != nil { | ||||||
|  | 		panic(err) | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // SchemeGroupVersion is the group version used to register these objects.
 | ||||||
|  | var SchemeGroupVersion = schema.GroupVersion{Group: zalando.GroupName, Version: APIVersion} | ||||||
|  | 
 | ||||||
|  | // Resource takes an unqualified resource and returns a Group-qualified GroupResource.
 | ||||||
|  | func Resource(resource string) schema.GroupResource { | ||||||
|  | 	return SchemeGroupVersion.WithResource(resource).GroupResource() | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // addKnownTypes adds the set of types defined in this package to the supplied scheme.
 | ||||||
|  | func addKnownTypes(scheme *runtime.Scheme) error { | ||||||
|  | 	scheme.AddKnownTypes(SchemeGroupVersion, | ||||||
|  | 		&FabricEventStream{}, | ||||||
|  | 		&FabricEventStreamList{}, | ||||||
|  | 	) | ||||||
|  | 	metav1.AddToGroupVersion(scheme, SchemeGroupVersion) | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | @ -0,0 +1,227 @@ | ||||||
|  | // +build !ignore_autogenerated
 | ||||||
|  | 
 | ||||||
|  | /* | ||||||
|  | Copyright 2021 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by deepcopy-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	runtime "k8s.io/apimachinery/pkg/runtime" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *Connection) DeepCopyInto(out *Connection) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *Connection) DeepCopy() *Connection { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(Connection) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *DBAuth) DeepCopyInto(out *DBAuth) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *DBAuth) DeepCopy() *DBAuth { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(DBAuth) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStream) DeepCopyInto(out *EventStream) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStream) DeepCopy() *EventStream { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(EventStream) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamFlow) DeepCopyInto(out *EventStreamFlow) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamFlow) DeepCopy() *EventStreamFlow { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(EventStreamFlow) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamSink) DeepCopyInto(out *EventStreamSink) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamSink) DeepCopy() *EventStreamSink { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(EventStreamSink) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamSource) DeepCopyInto(out *EventStreamSource) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamSource) DeepCopy() *EventStreamSource { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(EventStreamSource) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamTable) DeepCopyInto(out *EventStreamTable) { | ||||||
|  | 	*out = *in | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *EventStreamTable) DeepCopy() *EventStreamTable { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(EventStreamTable) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *FabricEventStreamSpec) DeepCopyInto(out *FabricEventStreamSpec) { | ||||||
|  | 	*out = *in | ||||||
|  | 	if in.EventStreams != nil { | ||||||
|  | 		in, out := &in.EventStreams, &out.EventStreams | ||||||
|  | 		*out = make([]EventStream, len(*in)) | ||||||
|  | 		for i := range *in { | ||||||
|  | 			(*in)[i].DeepCopyInto(&(*out)[i]) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FabricEventStreamSpec.
 | ||||||
|  | func (in *FabricEventStreamSpec) DeepCopy() *FabricEventStreamSpec { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(FabricEventStreamSpec) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *FabricEventStream) DeepCopyInto(out *FabricEventStream) { | ||||||
|  | 	*out = *in | ||||||
|  | 	out.TypeMeta = in.TypeMeta | ||||||
|  | 	in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) | ||||||
|  | 	in.Spec.DeepCopyInto(&out.Spec) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FabricEventStream.
 | ||||||
|  | func (in *FabricEventStream) DeepCopy() *FabricEventStream { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(FabricEventStream) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
 | ||||||
|  | func (in *FabricEventStream) DeepCopyObject() runtime.Object { | ||||||
|  | 	if c := in.DeepCopy(); c != nil { | ||||||
|  | 		return c | ||||||
|  | 	} | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
 | ||||||
|  | func (in *FabricEventStreamList) DeepCopyInto(out *FabricEventStreamList) { | ||||||
|  | 	*out = *in | ||||||
|  | 	out.TypeMeta = in.TypeMeta | ||||||
|  | 	in.ListMeta.DeepCopyInto(&out.ListMeta) | ||||||
|  | 	if in.Items != nil { | ||||||
|  | 		in, out := &in.Items, &out.Items | ||||||
|  | 		*out = make([]FabricEventStream, len(*in)) | ||||||
|  | 		for i := range *in { | ||||||
|  | 			(*in)[i].DeepCopyInto(&(*out)[i]) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new FabricEventStreamList.
 | ||||||
|  | func (in *FabricEventStreamList) DeepCopy() *FabricEventStreamList { | ||||||
|  | 	if in == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	out := new(FabricEventStreamList) | ||||||
|  | 	in.DeepCopyInto(out) | ||||||
|  | 	return out | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
 | ||||||
|  | func (in *FabricEventStreamList) DeepCopyObject() runtime.Object { | ||||||
|  | 	if c := in.DeepCopy(); c != nil { | ||||||
|  | 		return c | ||||||
|  | 	} | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | @ -361,6 +361,12 @@ func (c *Cluster) Create() error { | ||||||
| 	// something fails, report warning
 | 	// something fails, report warning
 | ||||||
| 	c.createConnectionPooler(c.installLookupFunction) | 	c.createConnectionPooler(c.installLookupFunction) | ||||||
| 
 | 
 | ||||||
|  | 	if len(c.Spec.Streams) > 0 { | ||||||
|  | 		if err = c.syncStreams(); err != nil { | ||||||
|  | 			c.logger.Errorf("could not create streams: %v", err) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
| 	return nil | 	return nil | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | @ -861,6 +867,13 @@ func (c *Cluster) Update(oldSpec, newSpec *acidv1.Postgresql) error { | ||||||
| 		updateFailed = true | 		updateFailed = true | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
|  | 	if len(c.Spec.Streams) > 0 { | ||||||
|  | 		if err := c.syncStreams(); err != nil { | ||||||
|  | 			c.logger.Errorf("could not sync streams: %v", err) | ||||||
|  | 			updateFailed = true | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
| 	if !updateFailed { | 	if !updateFailed { | ||||||
| 		// Major version upgrade must only fire after success of earlier operations and should stay last
 | 		// Major version upgrade must only fire after success of earlier operations and should stay last
 | ||||||
| 		if err := c.majorVersionUpgrade(); err != nil { | 		if err := c.majorVersionUpgrade(); err != nil { | ||||||
|  | @ -896,6 +909,10 @@ func (c *Cluster) Delete() { | ||||||
| 	defer c.mu.Unlock() | 	defer c.mu.Unlock() | ||||||
| 	c.eventRecorder.Event(c.GetReference(), v1.EventTypeNormal, "Delete", "Started deletion of new cluster resources") | 	c.eventRecorder.Event(c.GetReference(), v1.EventTypeNormal, "Delete", "Started deletion of new cluster resources") | ||||||
| 
 | 
 | ||||||
|  | 	if err := c.deleteStreams(); err != nil { | ||||||
|  | 		c.logger.Warningf("could not delete event streams: %v", err) | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
| 	// delete the backup job before the stateful set of the cluster to prevent connections to non-existing pods
 | 	// delete the backup job before the stateful set of the cluster to prevent connections to non-existing pods
 | ||||||
| 	// deleting the cron job also removes pods and batch jobs it created
 | 	// deleting the cron job also removes pods and batch jobs it created
 | ||||||
| 	if err := c.deleteLogicalBackupJob(); err != nil { | 	if err := c.deleteLogicalBackupJob(); err != nil { | ||||||
|  | @ -1050,6 +1067,23 @@ func (c *Cluster) initSystemUsers() { | ||||||
| 			c.systemUsers[constants.ConnectionPoolerUserKeyName] = connectionPoolerUser | 			c.systemUsers[constants.ConnectionPoolerUserKeyName] = connectionPoolerUser | ||||||
| 		} | 		} | ||||||
| 	} | 	} | ||||||
|  | 
 | ||||||
|  | 	// replication users for event streams are another exception
 | ||||||
|  | 	// the operator will create one replication user for all streams
 | ||||||
|  | 	if len(c.Spec.Streams) > 0 { | ||||||
|  | 		username := constants.EventStreamSourceSlotPrefix + constants.UserRoleNameSuffix | ||||||
|  | 		streamUser := spec.PgUser{ | ||||||
|  | 			Origin:    spec.RoleConnectionPooler, | ||||||
|  | 			Name:      username, | ||||||
|  | 			Namespace: c.Namespace, | ||||||
|  | 			Flags:     []string{constants.RoleFlagLogin, constants.RoleFlagReplication}, | ||||||
|  | 			Password:  util.RandomPassword(constants.PasswordLength), | ||||||
|  | 		} | ||||||
|  | 
 | ||||||
|  | 		if _, exists := c.pgUsers[username]; !exists { | ||||||
|  | 			c.pgUsers[username] = streamUser | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| func (c *Cluster) initPreparedDatabaseRoles() error { | func (c *Cluster) initPreparedDatabaseRoles() error { | ||||||
|  |  | ||||||
|  | @ -46,6 +46,13 @@ const ( | ||||||
| 	createExtensionSQL      = `CREATE EXTENSION IF NOT EXISTS "%s" SCHEMA "%s"` | 	createExtensionSQL      = `CREATE EXTENSION IF NOT EXISTS "%s" SCHEMA "%s"` | ||||||
| 	alterExtensionSQL       = `ALTER EXTENSION "%s" SET SCHEMA "%s"` | 	alterExtensionSQL       = `ALTER EXTENSION "%s" SET SCHEMA "%s"` | ||||||
| 
 | 
 | ||||||
|  | 	getPublicationsSQL = `SELECT p.pubname, string_agg(pt.schemaname || '.' || pt.tablename, ', ' ORDER BY pt.schemaname, pt.tablename) | ||||||
|  | 	        FROM pg_publication p | ||||||
|  | 			JOIN pg_publication_tables pt ON pt.pubname = p.pubname | ||||||
|  | 			GROUP BY p.pubname;` | ||||||
|  | 	createPublicationSQL = `CREATE PUBLICATION "%s" FOR TABLE %s WITH (publish = 'insert, update');` | ||||||
|  | 	alterPublicationSQL  = `ALTER PUBLICATION "%s" SET TABLE %s;` | ||||||
|  | 
 | ||||||
| 	globalDefaultPrivilegesSQL = `SET ROLE TO "%s"; | 	globalDefaultPrivilegesSQL = `SET ROLE TO "%s"; | ||||||
| 			ALTER DEFAULT PRIVILEGES GRANT USAGE ON SCHEMAS TO "%s","%s"; | 			ALTER DEFAULT PRIVILEGES GRANT USAGE ON SCHEMAS TO "%s","%s"; | ||||||
| 			ALTER DEFAULT PRIVILEGES GRANT SELECT ON TABLES TO "%s"; | 			ALTER DEFAULT PRIVILEGES GRANT SELECT ON TABLES TO "%s"; | ||||||
|  | @ -572,6 +579,68 @@ func (c *Cluster) execCreateOrAlterExtension(extName, schemaName, statement, doi | ||||||
| 	return nil | 	return nil | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | // getPublications returns the list of current database publications with tables
 | ||||||
|  | // The caller is responsible for opening and closing the database connection
 | ||||||
|  | func (c *Cluster) getPublications() (publications map[string]string, err error) { | ||||||
|  | 	var ( | ||||||
|  | 		rows *sql.Rows | ||||||
|  | 	) | ||||||
|  | 
 | ||||||
|  | 	if rows, err = c.pgDb.Query(getPublicationsSQL); err != nil { | ||||||
|  | 		return nil, fmt.Errorf("could not query database publications: %v", err) | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	defer func() { | ||||||
|  | 		if err2 := rows.Close(); err2 != nil { | ||||||
|  | 			if err != nil { | ||||||
|  | 				err = fmt.Errorf("error when closing query cursor: %v, previous error: %v", err2, err) | ||||||
|  | 			} else { | ||||||
|  | 				err = fmt.Errorf("error when closing query cursor: %v", err2) | ||||||
|  | 			} | ||||||
|  | 		} | ||||||
|  | 	}() | ||||||
|  | 
 | ||||||
|  | 	dbPublications := make(map[string]string) | ||||||
|  | 
 | ||||||
|  | 	for rows.Next() { | ||||||
|  | 		var ( | ||||||
|  | 			dbPublication       string | ||||||
|  | 			dbPublicationTables string | ||||||
|  | 		) | ||||||
|  | 
 | ||||||
|  | 		if err = rows.Scan(&dbPublication, &dbPublicationTables); err != nil { | ||||||
|  | 			return nil, fmt.Errorf("error when processing row: %v", err) | ||||||
|  | 		} | ||||||
|  | 		dbPublications[dbPublication] = dbPublicationTables | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return dbPublications, err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // executeCreatePublication creates new publication for given tables
 | ||||||
|  | // The caller is responsible for opening and closing the database connection.
 | ||||||
|  | func (c *Cluster) executeCreatePublication(pubName, tableList string) error { | ||||||
|  | 	return c.execCreateOrAlterPublication(pubName, tableList, createPublicationSQL, | ||||||
|  | 		"creating publication", "create publication") | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // executeAlterExtension changes the table list of the given publication.
 | ||||||
|  | // The caller is responsible for opening and closing the database connection.
 | ||||||
|  | func (c *Cluster) executeAlterPublication(pubName, tableList string) error { | ||||||
|  | 	return c.execCreateOrAlterPublication(pubName, tableList, alterPublicationSQL, | ||||||
|  | 		"changing publication", "alter publication tables") | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) execCreateOrAlterPublication(pubName, tableList, statement, doing, operation string) error { | ||||||
|  | 
 | ||||||
|  | 	c.logger.Debugf("%s %q with table list %q", doing, pubName, tableList) | ||||||
|  | 	if _, err := c.pgDb.Exec(fmt.Sprintf(statement, pubName, tableList)); err != nil { | ||||||
|  | 		return fmt.Errorf("could not execute %s: %v", operation, err) | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
| // Creates a connection pool credentials lookup function in every database to
 | // Creates a connection pool credentials lookup function in every database to
 | ||||||
| // perform remote authentication.
 | // perform remote authentication.
 | ||||||
| func (c *Cluster) installLookupFunction(poolerSchema, poolerUser string) error { | func (c *Cluster) installLookupFunction(poolerSchema, poolerUser string) error { | ||||||
|  |  | ||||||
|  | @ -0,0 +1,363 @@ | ||||||
|  | package cluster | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	"context" | ||||||
|  | 	"fmt" | ||||||
|  | 	"reflect" | ||||||
|  | 	"sort" | ||||||
|  | 	"strings" | ||||||
|  | 
 | ||||||
|  | 	acidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | ||||||
|  | 	zalandov1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/util" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/util/constants" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/util/k8sutil" | ||||||
|  | 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) createStreams(appId string) error { | ||||||
|  | 	c.setProcessName("creating streams") | ||||||
|  | 
 | ||||||
|  | 	fes := c.generateFabricEventStream(appId) | ||||||
|  | 	if _, err := c.KubeClient.FabricEventStreams(c.Namespace).Create(context.TODO(), fes, metav1.CreateOptions{}); err != nil { | ||||||
|  | 		return err | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) updateStreams(newEventStreams *zalandov1.FabricEventStream) error { | ||||||
|  | 	c.setProcessName("updating event streams") | ||||||
|  | 
 | ||||||
|  | 	if _, err := c.KubeClient.FabricEventStreams(newEventStreams.Namespace).Update(context.TODO(), newEventStreams, metav1.UpdateOptions{}); err != nil { | ||||||
|  | 		return err | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) deleteStreams() error { | ||||||
|  | 	c.setProcessName("deleting event streams") | ||||||
|  | 
 | ||||||
|  | 	// check if stream CRD is installed before trying a delete
 | ||||||
|  | 	_, err := c.KubeClient.CustomResourceDefinitions().Get(context.TODO(), constants.EventStreamCRDName, metav1.GetOptions{}) | ||||||
|  | 	if k8sutil.ResourceNotFound(err) { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	err = c.KubeClient.FabricEventStreams(c.Namespace).Delete(context.TODO(), c.Name, metav1.DeleteOptions{}) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return fmt.Errorf("could not delete event stream custom resource: %v", err) | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func gatherApplicationIds(streams []acidv1.Stream) []string { | ||||||
|  | 	appIds := make([]string, 0) | ||||||
|  | 	for _, stream := range streams { | ||||||
|  | 		if !util.SliceContains(appIds, stream.ApplicationId) { | ||||||
|  | 			appIds = append(appIds, stream.ApplicationId) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return appIds | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) syncPostgresConfig(requiredPatroniConfig acidv1.Patroni) error { | ||||||
|  | 	errorMsg := "no pods found to update config" | ||||||
|  | 
 | ||||||
|  | 	// if streams are defined wal_level must be switched to logical
 | ||||||
|  | 	requiredPgParameters := map[string]string{"wal_level": "logical"} | ||||||
|  | 
 | ||||||
|  | 	// apply config changes in pods
 | ||||||
|  | 	pods, err := c.listPods() | ||||||
|  | 	if err != nil { | ||||||
|  | 		errorMsg = fmt.Sprintf("could not list pods of the statefulset: %v", err) | ||||||
|  | 	} | ||||||
|  | 	for i, pod := range pods { | ||||||
|  | 		podName := util.NameFromMeta(pods[i].ObjectMeta) | ||||||
|  | 		effectivePatroniConfig, effectivePgParameters, err := c.patroni.GetConfig(&pod) | ||||||
|  | 		if err != nil { | ||||||
|  | 			errorMsg = fmt.Sprintf("could not get Postgres config from pod %s: %v", podName, err) | ||||||
|  | 			continue | ||||||
|  | 		} | ||||||
|  | 
 | ||||||
|  | 		_, err = c.checkAndSetGlobalPostgreSQLConfiguration(&pod, effectivePatroniConfig, requiredPatroniConfig, effectivePgParameters, requiredPgParameters) | ||||||
|  | 		if err != nil { | ||||||
|  | 			errorMsg = fmt.Sprintf("could not set PostgreSQL configuration options for pod %s: %v", podName, err) | ||||||
|  | 			continue | ||||||
|  | 		} | ||||||
|  | 
 | ||||||
|  | 		// Patroni's config endpoint is just a "proxy" to DCS. It is enough to patch it only once and it doesn't matter which pod is used
 | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return fmt.Errorf(errorMsg) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) syncPublication(publication, dbName string, tables map[string]acidv1.StreamTable) error { | ||||||
|  | 	createPublications := make(map[string]string) | ||||||
|  | 	alterPublications := make(map[string]string) | ||||||
|  | 
 | ||||||
|  | 	defer func() { | ||||||
|  | 		if err := c.closeDbConn(); err != nil { | ||||||
|  | 			c.logger.Errorf("could not close database connection: %v", err) | ||||||
|  | 		} | ||||||
|  | 	}() | ||||||
|  | 
 | ||||||
|  | 	// check for existing publications
 | ||||||
|  | 	if err := c.initDbConnWithName(dbName); err != nil { | ||||||
|  | 		return fmt.Errorf("could not init database connection") | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	currentPublications, err := c.getPublications() | ||||||
|  | 	if err != nil { | ||||||
|  | 		return fmt.Errorf("could not get current publications: %v", err) | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	tableNames := make([]string, len(tables)) | ||||||
|  | 	i := 0 | ||||||
|  | 	for t := range tables { | ||||||
|  | 		tableName, schemaName := getTableSchema(t) | ||||||
|  | 		tableNames[i] = fmt.Sprintf("%s.%s", schemaName, tableName) | ||||||
|  | 		i++ | ||||||
|  | 	} | ||||||
|  | 	sort.Strings(tableNames) | ||||||
|  | 	tableList := strings.Join(tableNames, ", ") | ||||||
|  | 
 | ||||||
|  | 	currentTables, exists := currentPublications[publication] | ||||||
|  | 	if !exists { | ||||||
|  | 		createPublications[publication] = tableList | ||||||
|  | 	} else if currentTables != tableList { | ||||||
|  | 		alterPublications[publication] = tableList | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	if len(createPublications)+len(alterPublications) == 0 { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	for publicationName, tables := range createPublications { | ||||||
|  | 		if err = c.executeCreatePublication(publicationName, tables); err != nil { | ||||||
|  | 			return fmt.Errorf("creation of publication %q failed: %v", publicationName, err) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 	for publicationName, tables := range alterPublications { | ||||||
|  | 		if err = c.executeAlterPublication(publicationName, tables); err != nil { | ||||||
|  | 			return fmt.Errorf("update of publication %q failed: %v", publicationName, err) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) generateFabricEventStream(appId string) *zalandov1.FabricEventStream { | ||||||
|  | 	eventStreams := make([]zalandov1.EventStream, 0) | ||||||
|  | 
 | ||||||
|  | 	for _, stream := range c.Spec.Streams { | ||||||
|  | 		if stream.ApplicationId != appId { | ||||||
|  | 			continue | ||||||
|  | 		} | ||||||
|  | 		for tableName, table := range stream.Tables { | ||||||
|  | 			streamSource := c.getEventStreamSource(stream, tableName, table.IdColumn) | ||||||
|  | 			streamFlow := getEventStreamFlow(stream, table.PayloadColumn) | ||||||
|  | 			streamSink := getEventStreamSink(stream, table.EventType) | ||||||
|  | 
 | ||||||
|  | 			eventStreams = append(eventStreams, zalandov1.EventStream{ | ||||||
|  | 				EventStreamFlow:   streamFlow, | ||||||
|  | 				EventStreamSink:   streamSink, | ||||||
|  | 				EventStreamSource: streamSource}) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return &zalandov1.FabricEventStream{ | ||||||
|  | 		TypeMeta: metav1.TypeMeta{ | ||||||
|  | 			APIVersion: constants.EventStreamCRDApiVersion, | ||||||
|  | 			Kind:       constants.EventStreamCRDKind, | ||||||
|  | 		}, | ||||||
|  | 		ObjectMeta: metav1.ObjectMeta{ | ||||||
|  | 			Name:        fmt.Sprintf("%s-%s", c.Name, appId), | ||||||
|  | 			Namespace:   c.Namespace, | ||||||
|  | 			Annotations: c.AnnotationsToPropagate(c.annotationsSet(nil)), | ||||||
|  | 			// make cluster StatefulSet the owner (like with connection pooler objects)
 | ||||||
|  | 			OwnerReferences: c.ownerReferences(), | ||||||
|  | 		}, | ||||||
|  | 		Spec: zalandov1.FabricEventStreamSpec{ | ||||||
|  | 			ApplicationId: appId, | ||||||
|  | 			EventStreams:  eventStreams, | ||||||
|  | 		}, | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) getEventStreamSource(stream acidv1.Stream, tableName, idColumn string) zalandov1.EventStreamSource { | ||||||
|  | 	table, schema := getTableSchema(tableName) | ||||||
|  | 	streamFilter := stream.Filter[tableName] | ||||||
|  | 	return zalandov1.EventStreamSource{ | ||||||
|  | 		Type:             constants.EventStreamSourcePGType, | ||||||
|  | 		Schema:           schema, | ||||||
|  | 		EventStreamTable: getOutboxTable(table, idColumn), | ||||||
|  | 		Filter:           streamFilter, | ||||||
|  | 		Connection: c.getStreamConnection( | ||||||
|  | 			stream.Database, | ||||||
|  | 			constants.EventStreamSourceSlotPrefix+constants.UserRoleNameSuffix, | ||||||
|  | 			stream.ApplicationId), | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func getEventStreamFlow(stream acidv1.Stream, payloadColumn string) zalandov1.EventStreamFlow { | ||||||
|  | 	return zalandov1.EventStreamFlow{ | ||||||
|  | 		Type:          constants.EventStreamFlowPgGenericType, | ||||||
|  | 		PayloadColumn: payloadColumn, | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func getEventStreamSink(stream acidv1.Stream, eventType string) zalandov1.EventStreamSink { | ||||||
|  | 	return zalandov1.EventStreamSink{ | ||||||
|  | 		Type:         constants.EventStreamSinkNakadiType, | ||||||
|  | 		EventType:    eventType, | ||||||
|  | 		MaxBatchSize: stream.BatchSize, | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func getTableSchema(fullTableName string) (tableName, schemaName string) { | ||||||
|  | 	schemaName = "public" | ||||||
|  | 	tableName = fullTableName | ||||||
|  | 	if strings.Contains(fullTableName, ".") { | ||||||
|  | 		schemaName = strings.Split(fullTableName, ".")[0] | ||||||
|  | 		tableName = strings.Split(fullTableName, ".")[1] | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return tableName, schemaName | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func getOutboxTable(tableName, idColumn string) zalandov1.EventStreamTable { | ||||||
|  | 	return zalandov1.EventStreamTable{ | ||||||
|  | 		Name:     tableName, | ||||||
|  | 		IDColumn: idColumn, | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func getSlotName(dbName, appId string) string { | ||||||
|  | 	return fmt.Sprintf("%s_%s_%s", constants.EventStreamSourceSlotPrefix, dbName, strings.Replace(appId, "-", "_", -1)) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) getStreamConnection(database, user, appId string) zalandov1.Connection { | ||||||
|  | 	return zalandov1.Connection{ | ||||||
|  | 		Url:        fmt.Sprintf("jdbc:postgresql://%s.%s/%s?user=%s&ssl=true&sslmode=require", c.Name, c.Namespace, database, user), | ||||||
|  | 		SlotName:   getSlotName(database, appId), | ||||||
|  | 		PluginType: constants.EventStreamSourcePluginType, | ||||||
|  | 		DBAuth: zalandov1.DBAuth{ | ||||||
|  | 			Type:        constants.EventStreamSourceAuthType, | ||||||
|  | 			Name:        c.credentialSecretNameForCluster(user, c.Name), | ||||||
|  | 			UserKey:     "username", | ||||||
|  | 			PasswordKey: "password", | ||||||
|  | 		}, | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) syncStreams() error { | ||||||
|  | 
 | ||||||
|  | 	c.setProcessName("syncing streams") | ||||||
|  | 
 | ||||||
|  | 	_, err := c.KubeClient.CustomResourceDefinitions().Get(context.TODO(), constants.EventStreamCRDName, metav1.GetOptions{}) | ||||||
|  | 	if k8sutil.ResourceNotFound(err) { | ||||||
|  | 		c.logger.Debugf("event stream CRD not installed, skipping") | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	slots := make(map[string]map[string]string) | ||||||
|  | 	publications := make(map[string]map[string]acidv1.StreamTable) | ||||||
|  | 
 | ||||||
|  | 	requiredPatroniConfig := c.Spec.Patroni | ||||||
|  | 	if len(requiredPatroniConfig.Slots) > 0 { | ||||||
|  | 		slots = requiredPatroniConfig.Slots | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	// gather list of required slots and publications
 | ||||||
|  | 	for _, stream := range c.Spec.Streams { | ||||||
|  | 		slot := map[string]string{ | ||||||
|  | 			"database": stream.Database, | ||||||
|  | 			"plugin":   constants.EventStreamSourcePluginType, | ||||||
|  | 			"type":     "logical", | ||||||
|  | 		} | ||||||
|  | 		slotName := getSlotName(stream.Database, stream.ApplicationId) | ||||||
|  | 		if _, exists := slots[slotName]; !exists { | ||||||
|  | 			slots[slotName] = slot | ||||||
|  | 			publications[slotName] = stream.Tables | ||||||
|  | 		} else { | ||||||
|  | 			streamTables := publications[slotName] | ||||||
|  | 			for tableName, table := range stream.Tables { | ||||||
|  | 				if _, exists := streamTables[tableName]; !exists { | ||||||
|  | 					streamTables[tableName] = table | ||||||
|  | 				} | ||||||
|  | 			} | ||||||
|  | 			publications[slotName] = streamTables | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	// no slots = no streams defined
 | ||||||
|  | 	if len(slots) > 0 { | ||||||
|  | 		requiredPatroniConfig.Slots = slots | ||||||
|  | 	} else { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	// add extra logical slots to Patroni config
 | ||||||
|  | 	c.logger.Debug("syncing Postgres config for logical decoding") | ||||||
|  | 	err = c.syncPostgresConfig(requiredPatroniConfig) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return fmt.Errorf("failed to snyc Postgres config for event streaming: %v", err) | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	// next, create publications to each created slot
 | ||||||
|  | 	c.logger.Debug("syncing database publications") | ||||||
|  | 	for publication, tables := range publications { | ||||||
|  | 		// but first check for existing publications
 | ||||||
|  | 		dbName := slots[publication]["database"] | ||||||
|  | 		err = c.syncPublication(publication, dbName, tables) | ||||||
|  | 		if err != nil { | ||||||
|  | 			c.logger.Warningf("could not sync publication %q in database %q: %v", publication, dbName, err) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	err = c.createOrUpdateStreams() | ||||||
|  | 	if err != nil { | ||||||
|  | 		return err | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *Cluster) createOrUpdateStreams() error { | ||||||
|  | 
 | ||||||
|  | 	appIds := gatherApplicationIds(c.Spec.Streams) | ||||||
|  | 	for _, appId := range appIds { | ||||||
|  | 		fesName := fmt.Sprintf("%s-%s", c.Name, appId) | ||||||
|  | 		effectiveStreams, err := c.KubeClient.FabricEventStreams(c.Namespace).Get(context.TODO(), fesName, metav1.GetOptions{}) | ||||||
|  | 		if err != nil { | ||||||
|  | 			if !k8sutil.ResourceNotFound(err) { | ||||||
|  | 				return fmt.Errorf("failed reading event stream %s: %v", fesName, err) | ||||||
|  | 			} | ||||||
|  | 
 | ||||||
|  | 			c.logger.Infof("event streams do not exist, create it") | ||||||
|  | 			err = c.createStreams(appId) | ||||||
|  | 			if err != nil { | ||||||
|  | 				return fmt.Errorf("failed creating event stream %s: %v", fesName, err) | ||||||
|  | 			} | ||||||
|  | 			c.logger.Infof("event stream %q has been successfully created", fesName) | ||||||
|  | 		} else { | ||||||
|  | 			desiredStreams := c.generateFabricEventStream(appId) | ||||||
|  | 			if !reflect.DeepEqual(effectiveStreams.Spec, desiredStreams.Spec) { | ||||||
|  | 				c.logger.Debug("updating event streams") | ||||||
|  | 				desiredStreams.ObjectMeta.ResourceVersion = effectiveStreams.ObjectMeta.ResourceVersion | ||||||
|  | 				err = c.updateStreams(desiredStreams) | ||||||
|  | 				if err != nil { | ||||||
|  | 					return fmt.Errorf("failed updating event stream %s: %v", fesName, err) | ||||||
|  | 				} | ||||||
|  | 				c.logger.Infof("event stream %q has been successfully updated", fesName) | ||||||
|  | 			} | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | @ -0,0 +1,244 @@ | ||||||
|  | package cluster | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	"encoding/json" | ||||||
|  | 	"fmt" | ||||||
|  | 	"reflect" | ||||||
|  | 	"strings" | ||||||
|  | 
 | ||||||
|  | 	"context" | ||||||
|  | 	"testing" | ||||||
|  | 
 | ||||||
|  | 	"github.com/stretchr/testify/assert" | ||||||
|  | 	acidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | ||||||
|  | 	v1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
|  | 	fakezalandov1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/fake" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/util" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/util/config" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/util/constants" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/util/k8sutil" | ||||||
|  | 
 | ||||||
|  | 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
|  | 	"k8s.io/apimachinery/pkg/types" | ||||||
|  | 	"k8s.io/client-go/kubernetes/fake" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | func newFakeK8sStreamClient() (k8sutil.KubernetesClient, *fake.Clientset) { | ||||||
|  | 	zalandoClientSet := fakezalandov1.NewSimpleClientset() | ||||||
|  | 	clientSet := fake.NewSimpleClientset() | ||||||
|  | 
 | ||||||
|  | 	return k8sutil.KubernetesClient{ | ||||||
|  | 		FabricEventStreamsGetter: zalandoClientSet.ZalandoV1(), | ||||||
|  | 		PostgresqlsGetter:        zalandoClientSet.AcidV1(), | ||||||
|  | 		PodsGetter:               clientSet.CoreV1(), | ||||||
|  | 		StatefulSetsGetter:       clientSet.AppsV1(), | ||||||
|  | 	}, clientSet | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | var ( | ||||||
|  | 	clusterName string = "acid-test-cluster" | ||||||
|  | 	namespace   string = "default" | ||||||
|  | 	appId       string = "test-app" | ||||||
|  | 	dbName      string = "foo" | ||||||
|  | 	fesUser     string = constants.EventStreamSourceSlotPrefix + constants.UserRoleNameSuffix | ||||||
|  | 	fesName     string = fmt.Sprintf("%s-%s", clusterName, appId) | ||||||
|  | 	slotName    string = fmt.Sprintf("%s_%s_%s", constants.EventStreamSourceSlotPrefix, dbName, strings.Replace(appId, "-", "_", -1)) | ||||||
|  | 
 | ||||||
|  | 	pg = acidv1.Postgresql{ | ||||||
|  | 		TypeMeta: metav1.TypeMeta{ | ||||||
|  | 			Kind:       "Postgresql", | ||||||
|  | 			APIVersion: "acid.zalan.do/v1", | ||||||
|  | 		}, | ||||||
|  | 		ObjectMeta: metav1.ObjectMeta{ | ||||||
|  | 			Name:      clusterName, | ||||||
|  | 			Namespace: namespace, | ||||||
|  | 		}, | ||||||
|  | 		Spec: acidv1.PostgresSpec{ | ||||||
|  | 			Databases: map[string]string{ | ||||||
|  | 				dbName: dbName + constants.UserRoleNameSuffix, | ||||||
|  | 			}, | ||||||
|  | 			Streams: []acidv1.Stream{ | ||||||
|  | 				{ | ||||||
|  | 					ApplicationId: appId, | ||||||
|  | 					Database:      "foo", | ||||||
|  | 					Tables: map[string]acidv1.StreamTable{ | ||||||
|  | 						"data.bar": acidv1.StreamTable{ | ||||||
|  | 							EventType:     "stream_type_a", | ||||||
|  | 							IdColumn:      "b_id", | ||||||
|  | 							PayloadColumn: "b_payload", | ||||||
|  | 						}, | ||||||
|  | 					}, | ||||||
|  | 					Filter: map[string]string{ | ||||||
|  | 						"data.bar": "[?(@.source.txId > 500 && @.source.lsn > 123456)]", | ||||||
|  | 					}, | ||||||
|  | 					BatchSize: uint32(100), | ||||||
|  | 				}, | ||||||
|  | 			}, | ||||||
|  | 			Volume: acidv1.Volume{ | ||||||
|  | 				Size: "1Gi", | ||||||
|  | 			}, | ||||||
|  | 		}, | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	fes = &v1.FabricEventStream{ | ||||||
|  | 		TypeMeta: metav1.TypeMeta{ | ||||||
|  | 			APIVersion: constants.EventStreamCRDApiVersion, | ||||||
|  | 			Kind:       constants.EventStreamCRDKind, | ||||||
|  | 		}, | ||||||
|  | 		ObjectMeta: metav1.ObjectMeta{ | ||||||
|  | 			Name:      fesName, | ||||||
|  | 			Namespace: namespace, | ||||||
|  | 			OwnerReferences: []metav1.OwnerReference{ | ||||||
|  | 				metav1.OwnerReference{ | ||||||
|  | 					APIVersion: "apps/v1", | ||||||
|  | 					Kind:       "StatefulSet", | ||||||
|  | 					Name:       "acid-test-cluster", | ||||||
|  | 					Controller: util.True(), | ||||||
|  | 				}, | ||||||
|  | 			}, | ||||||
|  | 		}, | ||||||
|  | 		Spec: v1.FabricEventStreamSpec{ | ||||||
|  | 			ApplicationId: appId, | ||||||
|  | 			EventStreams: []v1.EventStream{ | ||||||
|  | 				{ | ||||||
|  | 					EventStreamFlow: v1.EventStreamFlow{ | ||||||
|  | 						PayloadColumn: "b_payload", | ||||||
|  | 						Type:          constants.EventStreamFlowPgGenericType, | ||||||
|  | 					}, | ||||||
|  | 					EventStreamSink: v1.EventStreamSink{ | ||||||
|  | 						EventType:    "stream_type_a", | ||||||
|  | 						MaxBatchSize: uint32(100), | ||||||
|  | 						Type:         constants.EventStreamSinkNakadiType, | ||||||
|  | 					}, | ||||||
|  | 					EventStreamSource: v1.EventStreamSource{ | ||||||
|  | 						Filter: "[?(@.source.txId > 500 && @.source.lsn > 123456)]", | ||||||
|  | 						Connection: v1.Connection{ | ||||||
|  | 							DBAuth: v1.DBAuth{ | ||||||
|  | 								Name:        fmt.Sprintf("fes-user.%s.credentials.postgresql.acid.zalan.do", clusterName), | ||||||
|  | 								PasswordKey: "password", | ||||||
|  | 								Type:        constants.EventStreamSourceAuthType, | ||||||
|  | 								UserKey:     "username", | ||||||
|  | 							}, | ||||||
|  | 							Url:        fmt.Sprintf("jdbc:postgresql://%s.%s/foo?user=%s&ssl=true&sslmode=require", clusterName, namespace, fesUser), | ||||||
|  | 							SlotName:   slotName, | ||||||
|  | 							PluginType: constants.EventStreamSourcePluginType, | ||||||
|  | 						}, | ||||||
|  | 						Schema: "data", | ||||||
|  | 						EventStreamTable: v1.EventStreamTable{ | ||||||
|  | 							IDColumn: "b_id", | ||||||
|  | 							Name:     "bar", | ||||||
|  | 						}, | ||||||
|  | 						Type: constants.EventStreamSourcePGType, | ||||||
|  | 					}, | ||||||
|  | 				}, | ||||||
|  | 			}, | ||||||
|  | 		}, | ||||||
|  | 	} | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | func TestGenerateFabricEventStream(t *testing.T) { | ||||||
|  | 	client, _ := newFakeK8sStreamClient() | ||||||
|  | 
 | ||||||
|  | 	var cluster = New( | ||||||
|  | 		Config{ | ||||||
|  | 			OpConfig: config.Config{ | ||||||
|  | 				Auth: config.Auth{ | ||||||
|  | 					SecretNameTemplate: "{username}.{cluster}.credentials.{tprkind}.{tprgroup}", | ||||||
|  | 				}, | ||||||
|  | 				PodManagementPolicy: "ordered_ready", | ||||||
|  | 				Resources: config.Resources{ | ||||||
|  | 					ClusterLabels:        map[string]string{"application": "spilo"}, | ||||||
|  | 					ClusterNameLabel:     "cluster-name", | ||||||
|  | 					DefaultCPURequest:    "300m", | ||||||
|  | 					DefaultCPULimit:      "300m", | ||||||
|  | 					DefaultMemoryRequest: "300Mi", | ||||||
|  | 					DefaultMemoryLimit:   "300Mi", | ||||||
|  | 					PodRoleLabel:         "spilo-role", | ||||||
|  | 				}, | ||||||
|  | 			}, | ||||||
|  | 		}, client, pg, logger, eventRecorder) | ||||||
|  | 
 | ||||||
|  | 	cluster.Name = clusterName | ||||||
|  | 	cluster.Namespace = namespace | ||||||
|  | 
 | ||||||
|  | 	_, err := cluster.createStatefulSet() | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	err = cluster.createOrUpdateStreams() | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	result := cluster.generateFabricEventStream(appId) | ||||||
|  | 
 | ||||||
|  | 	if !reflect.DeepEqual(result, fes) { | ||||||
|  | 		t.Errorf("Malformed FabricEventStream, expected %#v, got %#v", fes, result) | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	streamCRD, err := cluster.KubeClient.FabricEventStreams(namespace).Get(context.TODO(), fesName, metav1.GetOptions{}) | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	if !reflect.DeepEqual(streamCRD, fes) { | ||||||
|  | 		t.Errorf("Malformed FabricEventStream, expected %#v, got %#v", fes, streamCRD) | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func TestUpdateFabricEventStream(t *testing.T) { | ||||||
|  | 	client, _ := newFakeK8sStreamClient() | ||||||
|  | 
 | ||||||
|  | 	var cluster = New( | ||||||
|  | 		Config{ | ||||||
|  | 			OpConfig: config.Config{ | ||||||
|  | 				PodManagementPolicy: "ordered_ready", | ||||||
|  | 				Resources: config.Resources{ | ||||||
|  | 					ClusterLabels:        map[string]string{"application": "spilo"}, | ||||||
|  | 					ClusterNameLabel:     "cluster-name", | ||||||
|  | 					DefaultCPURequest:    "300m", | ||||||
|  | 					DefaultCPULimit:      "300m", | ||||||
|  | 					DefaultMemoryRequest: "300Mi", | ||||||
|  | 					DefaultMemoryLimit:   "300Mi", | ||||||
|  | 					PodRoleLabel:         "spilo-role", | ||||||
|  | 				}, | ||||||
|  | 			}, | ||||||
|  | 		}, client, pg, logger, eventRecorder) | ||||||
|  | 
 | ||||||
|  | 	_, err := cluster.KubeClient.Postgresqls(namespace).Create( | ||||||
|  | 		context.TODO(), &pg, metav1.CreateOptions{}) | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 	err = cluster.createOrUpdateStreams() | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	var pgSpec acidv1.PostgresSpec | ||||||
|  | 	pgSpec.Streams = []acidv1.Stream{ | ||||||
|  | 		{ | ||||||
|  | 			ApplicationId: appId, | ||||||
|  | 			Database:      dbName, | ||||||
|  | 			Tables: map[string]acidv1.StreamTable{ | ||||||
|  | 				"data.bar": acidv1.StreamTable{ | ||||||
|  | 					EventType:     "stream_type_b", | ||||||
|  | 					IdColumn:      "b_id", | ||||||
|  | 					PayloadColumn: "b_payload", | ||||||
|  | 				}, | ||||||
|  | 			}, | ||||||
|  | 			BatchSize: uint32(250), | ||||||
|  | 		}, | ||||||
|  | 	} | ||||||
|  | 	patch, err := json.Marshal(struct { | ||||||
|  | 		PostgresSpec interface{} `json:"spec"` | ||||||
|  | 	}{&pgSpec}) | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	pgPatched, err := cluster.KubeClient.Postgresqls(namespace).Patch( | ||||||
|  | 		context.TODO(), cluster.Name, types.MergePatchType, patch, metav1.PatchOptions{}, "spec") | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	cluster.Postgresql.Spec = pgPatched.Spec | ||||||
|  | 	err = cluster.createOrUpdateStreams() | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	streamCRD, err := cluster.KubeClient.FabricEventStreams(namespace).Get(context.TODO(), fesName, metav1.GetOptions{}) | ||||||
|  | 	assert.NoError(t, err) | ||||||
|  | 
 | ||||||
|  | 	result := cluster.generateFabricEventStream(appId) | ||||||
|  | 	if !reflect.DeepEqual(result, streamCRD) { | ||||||
|  | 		t.Errorf("Malformed FabricEventStream, expected %#v, got %#v", streamCRD, result) | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | @ -81,7 +81,7 @@ func (c *Cluster) Sync(newSpec *acidv1.Postgresql) error { | ||||||
| 		return err | 		return err | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	c.logger.Debugf("syncing statefulsets") | 	c.logger.Debug("syncing statefulsets") | ||||||
| 	if err = c.syncStatefulSet(); err != nil { | 	if err = c.syncStatefulSet(); err != nil { | ||||||
| 		if !k8sutil.ResourceAlreadyExists(err) { | 		if !k8sutil.ResourceAlreadyExists(err) { | ||||||
| 			err = fmt.Errorf("could not sync statefulsets: %v", err) | 			err = fmt.Errorf("could not sync statefulsets: %v", err) | ||||||
|  | @ -107,17 +107,17 @@ func (c *Cluster) Sync(newSpec *acidv1.Postgresql) error { | ||||||
| 
 | 
 | ||||||
| 	// create database objects unless we are running without pods or disabled that feature explicitly
 | 	// create database objects unless we are running without pods or disabled that feature explicitly
 | ||||||
| 	if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&newSpec.Spec) <= 0 || c.Spec.StandbyCluster != nil) { | 	if !(c.databaseAccessDisabled() || c.getNumberOfInstances(&newSpec.Spec) <= 0 || c.Spec.StandbyCluster != nil) { | ||||||
| 		c.logger.Debugf("syncing roles") | 		c.logger.Debug("syncing roles") | ||||||
| 		if err = c.syncRoles(); err != nil { | 		if err = c.syncRoles(); err != nil { | ||||||
| 			err = fmt.Errorf("could not sync roles: %v", err) | 			err = fmt.Errorf("could not sync roles: %v", err) | ||||||
| 			return err | 			return err | ||||||
| 		} | 		} | ||||||
| 		c.logger.Debugf("syncing databases") | 		c.logger.Debug("syncing databases") | ||||||
| 		if err = c.syncDatabases(); err != nil { | 		if err = c.syncDatabases(); err != nil { | ||||||
| 			err = fmt.Errorf("could not sync databases: %v", err) | 			err = fmt.Errorf("could not sync databases: %v", err) | ||||||
| 			return err | 			return err | ||||||
| 		} | 		} | ||||||
| 		c.logger.Debugf("syncing prepared databases with schemas") | 		c.logger.Debug("syncing prepared databases with schemas") | ||||||
| 		if err = c.syncPreparedDatabases(); err != nil { | 		if err = c.syncPreparedDatabases(); err != nil { | ||||||
| 			err = fmt.Errorf("could not sync prepared database: %v", err) | 			err = fmt.Errorf("could not sync prepared database: %v", err) | ||||||
| 			return err | 			return err | ||||||
|  | @ -129,6 +129,14 @@ func (c *Cluster) Sync(newSpec *acidv1.Postgresql) error { | ||||||
| 		return fmt.Errorf("could not sync connection pooler: %v", err) | 		return fmt.Errorf("could not sync connection pooler: %v", err) | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
|  | 	if len(c.Spec.Streams) > 0 { | ||||||
|  | 		c.logger.Debug("syncing streams") | ||||||
|  | 		if err = c.syncStreams(); err != nil { | ||||||
|  | 			err = fmt.Errorf("could not sync streams: %v", err) | ||||||
|  | 			return err | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
| 	// Major version upgrade must only run after success of all earlier operations, must remain last item in sync
 | 	// Major version upgrade must only run after success of all earlier operations, must remain last item in sync
 | ||||||
| 	if err := c.majorVersionUpgrade(); err != nil { | 	if err := c.majorVersionUpgrade(); err != nil { | ||||||
| 		c.logger.Errorf("major version upgrade failed: %v", err) | 		c.logger.Errorf("major version upgrade failed: %v", err) | ||||||
|  | @ -401,7 +409,7 @@ func (c *Cluster) syncStatefulSet() error { | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	// get Postgres config, compare with manifest and update via Patroni PATCH endpoint if it differs
 | 	// get Postgres config, compare with manifest and update via Patroni PATCH endpoint if it differs
 | ||||||
| 	// Patroni's config endpoint is just a "proxy" to DCS. It is enough to patch it only once and it doesn't matter which pod is used.
 | 	// Patroni's config endpoint is just a "proxy" to DCS. It is enough to patch it only once and it doesn't matter which pod is used
 | ||||||
| 	for i, pod := range pods { | 	for i, pod := range pods { | ||||||
| 		patroniConfig, pgParameters, err := c.getPatroniConfig(&pod) | 		patroniConfig, pgParameters, err := c.getPatroniConfig(&pod) | ||||||
| 		if err != nil { | 		if err != nil { | ||||||
|  | @ -414,11 +422,13 @@ func (c *Cluster) syncStatefulSet() error { | ||||||
| 		// empty config probably means cluster is not fully initialized yet, e.g. restoring from backup
 | 		// empty config probably means cluster is not fully initialized yet, e.g. restoring from backup
 | ||||||
| 		// do not attempt a restart
 | 		// do not attempt a restart
 | ||||||
| 		if !reflect.DeepEqual(patroniConfig, acidv1.Patroni{}) || len(pgParameters) > 0 { | 		if !reflect.DeepEqual(patroniConfig, acidv1.Patroni{}) || len(pgParameters) > 0 { | ||||||
| 			restartMasterFirst, err = c.checkAndSetGlobalPostgreSQLConfiguration(&pod, patroniConfig, pgParameters) | 			// compare config returned from Patroni with what is specified in the manifest
 | ||||||
|  | 			restartMasterFirst, err = c.checkAndSetGlobalPostgreSQLConfiguration(&pod, patroniConfig, c.Spec.Patroni, pgParameters, c.Spec.Parameters) | ||||||
| 			if err != nil { | 			if err != nil { | ||||||
| 				c.logger.Warningf("could not set PostgreSQL configuration options for pod %s: %v", pods[i].Name, err) | 				c.logger.Warningf("could not set PostgreSQL configuration options for pod %s: %v", pods[i].Name, err) | ||||||
| 				continue | 				continue | ||||||
| 			} | 			} | ||||||
|  | 
 | ||||||
| 			// it could take up to LoopWait to apply the config
 | 			// it could take up to LoopWait to apply the config
 | ||||||
| 			time.Sleep(time.Duration(restartWait)*time.Second + time.Second*2) | 			time.Sleep(time.Duration(restartWait)*time.Second + time.Second*2) | ||||||
| 			break | 			break | ||||||
|  | @ -526,40 +536,39 @@ func (c *Cluster) AnnotationsToPropagate(annotations map[string]string) map[stri | ||||||
| 
 | 
 | ||||||
| // checkAndSetGlobalPostgreSQLConfiguration checks whether cluster-wide API parameters
 | // checkAndSetGlobalPostgreSQLConfiguration checks whether cluster-wide API parameters
 | ||||||
| // (like max_connections) have changed and if necessary sets it via the Patroni API
 | // (like max_connections) have changed and if necessary sets it via the Patroni API
 | ||||||
| func (c *Cluster) checkAndSetGlobalPostgreSQLConfiguration(pod *v1.Pod, patroniConfig acidv1.Patroni, effectivePgParameters map[string]string) (bool, error) { | func (c *Cluster) checkAndSetGlobalPostgreSQLConfiguration(pod *v1.Pod, effectivePatroniConfig, desiredPatroniConfig acidv1.Patroni, effectivePgParameters, desiredPgParameters map[string]string) (bool, error) { | ||||||
| 	configToSet := make(map[string]interface{}) | 	configToSet := make(map[string]interface{}) | ||||||
| 	parametersToSet := make(map[string]string) | 	parametersToSet := make(map[string]string) | ||||||
| 	restartMaster := make([]bool, 0) | 	restartMaster := make([]bool, 0) | ||||||
| 	requiresMasterRestart := false | 	requiresMasterRestart := false | ||||||
| 
 | 
 | ||||||
| 	// compare options from config with c.Spec.Patroni from manifest
 | 	// compare effective and desired Patroni config options
 | ||||||
| 	desiredPatroniConfig := c.Spec.Patroni | 	if desiredPatroniConfig.LoopWait > 0 && desiredPatroniConfig.LoopWait != effectivePatroniConfig.LoopWait { | ||||||
| 	if desiredPatroniConfig.LoopWait > 0 && desiredPatroniConfig.LoopWait != patroniConfig.LoopWait { |  | ||||||
| 		configToSet["loop_wait"] = desiredPatroniConfig.LoopWait | 		configToSet["loop_wait"] = desiredPatroniConfig.LoopWait | ||||||
| 	} | 	} | ||||||
| 	if desiredPatroniConfig.MaximumLagOnFailover > 0 && desiredPatroniConfig.MaximumLagOnFailover != patroniConfig.MaximumLagOnFailover { | 	if desiredPatroniConfig.MaximumLagOnFailover > 0 && desiredPatroniConfig.MaximumLagOnFailover != effectivePatroniConfig.MaximumLagOnFailover { | ||||||
| 		configToSet["maximum_lag_on_failover"] = desiredPatroniConfig.MaximumLagOnFailover | 		configToSet["maximum_lag_on_failover"] = desiredPatroniConfig.MaximumLagOnFailover | ||||||
| 	} | 	} | ||||||
| 	if desiredPatroniConfig.PgHba != nil && !reflect.DeepEqual(desiredPatroniConfig.PgHba, patroniConfig.PgHba) { | 	if desiredPatroniConfig.PgHba != nil && !reflect.DeepEqual(desiredPatroniConfig.PgHba, effectivePatroniConfig.PgHba) { | ||||||
| 		configToSet["pg_hba"] = desiredPatroniConfig.PgHba | 		configToSet["pg_hba"] = desiredPatroniConfig.PgHba | ||||||
| 	} | 	} | ||||||
| 	if desiredPatroniConfig.RetryTimeout > 0 && desiredPatroniConfig.RetryTimeout != patroniConfig.RetryTimeout { | 	if desiredPatroniConfig.RetryTimeout > 0 && desiredPatroniConfig.RetryTimeout != effectivePatroniConfig.RetryTimeout { | ||||||
| 		configToSet["retry_timeout"] = desiredPatroniConfig.RetryTimeout | 		configToSet["retry_timeout"] = desiredPatroniConfig.RetryTimeout | ||||||
| 	} | 	} | ||||||
| 	if desiredPatroniConfig.SynchronousMode != patroniConfig.SynchronousMode { | 	if desiredPatroniConfig.SynchronousMode != effectivePatroniConfig.SynchronousMode { | ||||||
| 		configToSet["synchronous_mode"] = desiredPatroniConfig.SynchronousMode | 		configToSet["synchronous_mode"] = desiredPatroniConfig.SynchronousMode | ||||||
| 	} | 	} | ||||||
| 	if desiredPatroniConfig.SynchronousModeStrict != patroniConfig.SynchronousModeStrict { | 	if desiredPatroniConfig.SynchronousModeStrict != effectivePatroniConfig.SynchronousModeStrict { | ||||||
| 		configToSet["synchronous_mode_strict"] = desiredPatroniConfig.SynchronousModeStrict | 		configToSet["synchronous_mode_strict"] = desiredPatroniConfig.SynchronousModeStrict | ||||||
| 	} | 	} | ||||||
| 	if desiredPatroniConfig.TTL > 0 && desiredPatroniConfig.TTL != patroniConfig.TTL { | 	if desiredPatroniConfig.TTL > 0 && desiredPatroniConfig.TTL != effectivePatroniConfig.TTL { | ||||||
| 		configToSet["ttl"] = desiredPatroniConfig.TTL | 		configToSet["ttl"] = desiredPatroniConfig.TTL | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	// check if specified slots exist in config and if they differ
 | 	// check if specified slots exist in config and if they differ
 | ||||||
| 	slotsToSet := make(map[string]map[string]string) | 	slotsToSet := make(map[string]map[string]string) | ||||||
| 	for slotName, desiredSlot := range desiredPatroniConfig.Slots { | 	for slotName, desiredSlot := range desiredPatroniConfig.Slots { | ||||||
| 		if effectiveSlot, exists := patroniConfig.Slots[slotName]; exists { | 		if effectiveSlot, exists := effectivePatroniConfig.Slots[slotName]; exists { | ||||||
| 			if reflect.DeepEqual(desiredSlot, effectiveSlot) { | 			if reflect.DeepEqual(desiredSlot, effectiveSlot) { | ||||||
| 				continue | 				continue | ||||||
| 			} | 			} | ||||||
|  | @ -570,8 +579,7 @@ func (c *Cluster) checkAndSetGlobalPostgreSQLConfiguration(pod *v1.Pod, patroniC | ||||||
| 		configToSet["slots"] = slotsToSet | 		configToSet["slots"] = slotsToSet | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	// compare parameters under postgresql section with c.Spec.Postgresql.Parameters from manifest
 | 	// compare effective and desired parameters under postgresql section in Patroni config
 | ||||||
| 	desiredPgParameters := c.Spec.Parameters |  | ||||||
| 	for desiredOption, desiredValue := range desiredPgParameters { | 	for desiredOption, desiredValue := range desiredPgParameters { | ||||||
| 		effectiveValue := effectivePgParameters[desiredOption] | 		effectiveValue := effectivePgParameters[desiredOption] | ||||||
| 		if isBootstrapOnlyParameter(desiredOption) && (effectiveValue != desiredValue) { | 		if isBootstrapOnlyParameter(desiredOption) && (effectiveValue != desiredValue) { | ||||||
|  | @ -614,7 +622,7 @@ func (c *Cluster) checkAndSetGlobalPostgreSQLConfiguration(pod *v1.Pod, patroniC | ||||||
| 	c.logger.Debugf("patching Postgres config via Patroni API on pod %s with following options: %s", | 	c.logger.Debugf("patching Postgres config via Patroni API on pod %s with following options: %s", | ||||||
| 		podName, configToSetJson) | 		podName, configToSetJson) | ||||||
| 	if err = c.patroni.SetConfig(pod, configToSet); err != nil { | 	if err = c.patroni.SetConfig(pod, configToSet); err != nil { | ||||||
| 		return requiresMasterRestart, fmt.Errorf("could not patch postgres parameters with a pod %s: %v", podName, err) | 		return requiresMasterRestart, fmt.Errorf("could not patch postgres parameters within pod %s: %v", podName, err) | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	return requiresMasterRestart, nil | 	return requiresMasterRestart, nil | ||||||
|  |  | ||||||
|  | @ -256,7 +256,7 @@ func TestCheckAndSetGlobalPostgreSQLConfiguration(t *testing.T) { | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	for _, tt := range tests { | 	for _, tt := range tests { | ||||||
| 		requireMasterRestart, err := cluster.checkAndSetGlobalPostgreSQLConfiguration(mockPod, tt.patroni, tt.pgParams) | 		requireMasterRestart, err := cluster.checkAndSetGlobalPostgreSQLConfiguration(mockPod, tt.patroni, cluster.Spec.Patroni, tt.pgParams, cluster.Spec.Parameters) | ||||||
| 		assert.NoError(t, err) | 		assert.NoError(t, err) | ||||||
| 		if requireMasterRestart != tt.restartMaster { | 		if requireMasterRestart != tt.restartMaster { | ||||||
| 			t.Errorf("%s - %s: unexpect master restart strategy, got %v, expected %v", testName, tt.subtest, requireMasterRestart, tt.restartMaster) | 			t.Errorf("%s - %s: unexpect master restart strategy, got %v, expected %v", testName, tt.subtest, requireMasterRestart, tt.restartMaster) | ||||||
|  |  | ||||||
|  | @ -28,6 +28,7 @@ import ( | ||||||
| 	"fmt" | 	"fmt" | ||||||
| 
 | 
 | ||||||
| 	acidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1" | 	acidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1" | ||||||
|  | 	zalandov1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/zalando.org/v1" | ||||||
| 	discovery "k8s.io/client-go/discovery" | 	discovery "k8s.io/client-go/discovery" | ||||||
| 	rest "k8s.io/client-go/rest" | 	rest "k8s.io/client-go/rest" | ||||||
| 	flowcontrol "k8s.io/client-go/util/flowcontrol" | 	flowcontrol "k8s.io/client-go/util/flowcontrol" | ||||||
|  | @ -36,13 +37,15 @@ import ( | ||||||
| type Interface interface { | type Interface interface { | ||||||
| 	Discovery() discovery.DiscoveryInterface | 	Discovery() discovery.DiscoveryInterface | ||||||
| 	AcidV1() acidv1.AcidV1Interface | 	AcidV1() acidv1.AcidV1Interface | ||||||
|  | 	ZalandoV1() zalandov1.ZalandoV1Interface | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| // Clientset contains the clients for groups. Each group has exactly one
 | // Clientset contains the clients for groups. Each group has exactly one
 | ||||||
| // version included in a Clientset.
 | // version included in a Clientset.
 | ||||||
| type Clientset struct { | type Clientset struct { | ||||||
| 	*discovery.DiscoveryClient | 	*discovery.DiscoveryClient | ||||||
| 	acidV1 *acidv1.AcidV1Client | 	acidV1    *acidv1.AcidV1Client | ||||||
|  | 	zalandoV1 *zalandov1.ZalandoV1Client | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| // AcidV1 retrieves the AcidV1Client
 | // AcidV1 retrieves the AcidV1Client
 | ||||||
|  | @ -50,6 +53,11 @@ func (c *Clientset) AcidV1() acidv1.AcidV1Interface { | ||||||
| 	return c.acidV1 | 	return c.acidV1 | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | // ZalandoV1 retrieves the ZalandoV1Client
 | ||||||
|  | func (c *Clientset) ZalandoV1() zalandov1.ZalandoV1Interface { | ||||||
|  | 	return c.zalandoV1 | ||||||
|  | } | ||||||
|  | 
 | ||||||
| // Discovery retrieves the DiscoveryClient
 | // Discovery retrieves the DiscoveryClient
 | ||||||
| func (c *Clientset) Discovery() discovery.DiscoveryInterface { | func (c *Clientset) Discovery() discovery.DiscoveryInterface { | ||||||
| 	if c == nil { | 	if c == nil { | ||||||
|  | @ -75,6 +83,10 @@ func NewForConfig(c *rest.Config) (*Clientset, error) { | ||||||
| 	if err != nil { | 	if err != nil { | ||||||
| 		return nil, err | 		return nil, err | ||||||
| 	} | 	} | ||||||
|  | 	cs.zalandoV1, err = zalandov1.NewForConfig(&configShallowCopy) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
| 
 | 
 | ||||||
| 	cs.DiscoveryClient, err = discovery.NewDiscoveryClientForConfig(&configShallowCopy) | 	cs.DiscoveryClient, err = discovery.NewDiscoveryClientForConfig(&configShallowCopy) | ||||||
| 	if err != nil { | 	if err != nil { | ||||||
|  | @ -88,6 +100,7 @@ func NewForConfig(c *rest.Config) (*Clientset, error) { | ||||||
| func NewForConfigOrDie(c *rest.Config) *Clientset { | func NewForConfigOrDie(c *rest.Config) *Clientset { | ||||||
| 	var cs Clientset | 	var cs Clientset | ||||||
| 	cs.acidV1 = acidv1.NewForConfigOrDie(c) | 	cs.acidV1 = acidv1.NewForConfigOrDie(c) | ||||||
|  | 	cs.zalandoV1 = zalandov1.NewForConfigOrDie(c) | ||||||
| 
 | 
 | ||||||
| 	cs.DiscoveryClient = discovery.NewDiscoveryClientForConfigOrDie(c) | 	cs.DiscoveryClient = discovery.NewDiscoveryClientForConfigOrDie(c) | ||||||
| 	return &cs | 	return &cs | ||||||
|  | @ -97,6 +110,7 @@ func NewForConfigOrDie(c *rest.Config) *Clientset { | ||||||
| func New(c rest.Interface) *Clientset { | func New(c rest.Interface) *Clientset { | ||||||
| 	var cs Clientset | 	var cs Clientset | ||||||
| 	cs.acidV1 = acidv1.New(c) | 	cs.acidV1 = acidv1.New(c) | ||||||
|  | 	cs.zalandoV1 = zalandov1.New(c) | ||||||
| 
 | 
 | ||||||
| 	cs.DiscoveryClient = discovery.NewDiscoveryClient(c) | 	cs.DiscoveryClient = discovery.NewDiscoveryClient(c) | ||||||
| 	return &cs | 	return &cs | ||||||
|  |  | ||||||
|  | @ -28,6 +28,8 @@ import ( | ||||||
| 	clientset "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned" | 	clientset "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned" | ||||||
| 	acidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1" | 	acidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1" | ||||||
| 	fakeacidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1/fake" | 	fakeacidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1/fake" | ||||||
|  | 	zalandov1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/zalando.org/v1" | ||||||
|  | 	fakezalandov1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/zalando.org/v1/fake" | ||||||
| 	"k8s.io/apimachinery/pkg/runtime" | 	"k8s.io/apimachinery/pkg/runtime" | ||||||
| 	"k8s.io/apimachinery/pkg/watch" | 	"k8s.io/apimachinery/pkg/watch" | ||||||
| 	"k8s.io/client-go/discovery" | 	"k8s.io/client-go/discovery" | ||||||
|  | @ -89,3 +91,8 @@ var ( | ||||||
| func (c *Clientset) AcidV1() acidv1.AcidV1Interface { | func (c *Clientset) AcidV1() acidv1.AcidV1Interface { | ||||||
| 	return &fakeacidv1.FakeAcidV1{Fake: &c.Fake} | 	return &fakeacidv1.FakeAcidV1{Fake: &c.Fake} | ||||||
| } | } | ||||||
|  | 
 | ||||||
|  | // ZalandoV1 retrieves the ZalandoV1Client
 | ||||||
|  | func (c *Clientset) ZalandoV1() zalandov1.ZalandoV1Interface { | ||||||
|  | 	return &fakezalandov1.FakeZalandoV1{Fake: &c.Fake} | ||||||
|  | } | ||||||
|  |  | ||||||
|  | @ -26,6 +26,7 @@ package fake | ||||||
| 
 | 
 | ||||||
| import ( | import ( | ||||||
| 	acidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | 	acidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | ||||||
|  | 	zalandov1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
| 	v1 "k8s.io/apimachinery/pkg/apis/meta/v1" | 	v1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
| 	runtime "k8s.io/apimachinery/pkg/runtime" | 	runtime "k8s.io/apimachinery/pkg/runtime" | ||||||
| 	schema "k8s.io/apimachinery/pkg/runtime/schema" | 	schema "k8s.io/apimachinery/pkg/runtime/schema" | ||||||
|  | @ -38,6 +39,7 @@ var codecs = serializer.NewCodecFactory(scheme) | ||||||
| 
 | 
 | ||||||
| var localSchemeBuilder = runtime.SchemeBuilder{ | var localSchemeBuilder = runtime.SchemeBuilder{ | ||||||
| 	acidv1.AddToScheme, | 	acidv1.AddToScheme, | ||||||
|  | 	zalandov1.AddToScheme, | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| // AddToScheme adds all types of this clientset into the given scheme. This allows composition
 | // AddToScheme adds all types of this clientset into the given scheme. This allows composition
 | ||||||
|  |  | ||||||
|  | @ -26,6 +26,7 @@ package scheme | ||||||
| 
 | 
 | ||||||
| import ( | import ( | ||||||
| 	acidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | 	acidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | ||||||
|  | 	zalandov1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
| 	v1 "k8s.io/apimachinery/pkg/apis/meta/v1" | 	v1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
| 	runtime "k8s.io/apimachinery/pkg/runtime" | 	runtime "k8s.io/apimachinery/pkg/runtime" | ||||||
| 	schema "k8s.io/apimachinery/pkg/runtime/schema" | 	schema "k8s.io/apimachinery/pkg/runtime/schema" | ||||||
|  | @ -38,6 +39,7 @@ var Codecs = serializer.NewCodecFactory(Scheme) | ||||||
| var ParameterCodec = runtime.NewParameterCodec(Scheme) | var ParameterCodec = runtime.NewParameterCodec(Scheme) | ||||||
| var localSchemeBuilder = runtime.SchemeBuilder{ | var localSchemeBuilder = runtime.SchemeBuilder{ | ||||||
| 	acidv1.AddToScheme, | 	acidv1.AddToScheme, | ||||||
|  | 	zalandov1.AddToScheme, | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| // AddToScheme adds all types of this clientset into the given scheme. This allows composition
 | // AddToScheme adds all types of this clientset into the given scheme. This allows composition
 | ||||||
|  |  | ||||||
|  | @ -0,0 +1,26 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by client-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | // This package has the automatically generated typed clients.
 | ||||||
|  | package v1 | ||||||
|  | @ -0,0 +1,184 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by client-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	"context" | ||||||
|  | 	"time" | ||||||
|  | 
 | ||||||
|  | 	v1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
|  | 	scheme "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/scheme" | ||||||
|  | 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
|  | 	types "k8s.io/apimachinery/pkg/types" | ||||||
|  | 	watch "k8s.io/apimachinery/pkg/watch" | ||||||
|  | 	rest "k8s.io/client-go/rest" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamsGetter has a method to return a FabricEventStreamInterface.
 | ||||||
|  | // A group's client should implement this interface.
 | ||||||
|  | type FabricEventStreamsGetter interface { | ||||||
|  | 	FabricEventStreams(namespace string) FabricEventStreamInterface | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamInterface has methods to work with FabricEventStream resources.
 | ||||||
|  | type FabricEventStreamInterface interface { | ||||||
|  | 	Create(ctx context.Context, fabricEventStream *v1.FabricEventStream, opts metav1.CreateOptions) (*v1.FabricEventStream, error) | ||||||
|  | 	Update(ctx context.Context, fabricEventStream *v1.FabricEventStream, opts metav1.UpdateOptions) (*v1.FabricEventStream, error) | ||||||
|  | 	Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error | ||||||
|  | 	DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error | ||||||
|  | 	Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.FabricEventStream, error) | ||||||
|  | 	List(ctx context.Context, opts metav1.ListOptions) (*v1.FabricEventStreamList, error) | ||||||
|  | 	Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) | ||||||
|  | 	Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.FabricEventStream, err error) | ||||||
|  | 	FabricEventStreamExpansion | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // fabricEventStreams implements FabricEventStreamInterface
 | ||||||
|  | type fabricEventStreams struct { | ||||||
|  | 	client rest.Interface | ||||||
|  | 	ns     string | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // newFabricEventStreams returns a FabricEventStreams
 | ||||||
|  | func newFabricEventStreams(c *ZalandoV1Client, namespace string) *fabricEventStreams { | ||||||
|  | 	return &fabricEventStreams{ | ||||||
|  | 		client: c.RESTClient(), | ||||||
|  | 		ns:     namespace, | ||||||
|  | 	} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Get takes name of the fabricEventStream, and returns the corresponding fabricEventStream object, and an error if there is any.
 | ||||||
|  | func (c *fabricEventStreams) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.FabricEventStream, err error) { | ||||||
|  | 	result = &v1.FabricEventStream{} | ||||||
|  | 	err = c.client.Get(). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		Name(name). | ||||||
|  | 		VersionedParams(&options, scheme.ParameterCodec). | ||||||
|  | 		Do(ctx). | ||||||
|  | 		Into(result) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // List takes label and field selectors, and returns the list of FabricEventStreams that match those selectors.
 | ||||||
|  | func (c *fabricEventStreams) List(ctx context.Context, opts metav1.ListOptions) (result *v1.FabricEventStreamList, err error) { | ||||||
|  | 	var timeout time.Duration | ||||||
|  | 	if opts.TimeoutSeconds != nil { | ||||||
|  | 		timeout = time.Duration(*opts.TimeoutSeconds) * time.Second | ||||||
|  | 	} | ||||||
|  | 	result = &v1.FabricEventStreamList{} | ||||||
|  | 	err = c.client.Get(). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		VersionedParams(&opts, scheme.ParameterCodec). | ||||||
|  | 		Timeout(timeout). | ||||||
|  | 		Do(ctx). | ||||||
|  | 		Into(result) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Watch returns a watch.Interface that watches the requested fabricEventStreams.
 | ||||||
|  | func (c *fabricEventStreams) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { | ||||||
|  | 	var timeout time.Duration | ||||||
|  | 	if opts.TimeoutSeconds != nil { | ||||||
|  | 		timeout = time.Duration(*opts.TimeoutSeconds) * time.Second | ||||||
|  | 	} | ||||||
|  | 	opts.Watch = true | ||||||
|  | 	return c.client.Get(). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		VersionedParams(&opts, scheme.ParameterCodec). | ||||||
|  | 		Timeout(timeout). | ||||||
|  | 		Watch(ctx) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Create takes the representation of a fabricEventStream and creates it.  Returns the server's representation of the fabricEventStream, and an error, if there is any.
 | ||||||
|  | func (c *fabricEventStreams) Create(ctx context.Context, fabricEventStream *v1.FabricEventStream, opts metav1.CreateOptions) (result *v1.FabricEventStream, err error) { | ||||||
|  | 	result = &v1.FabricEventStream{} | ||||||
|  | 	err = c.client.Post(). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		VersionedParams(&opts, scheme.ParameterCodec). | ||||||
|  | 		Body(fabricEventStream). | ||||||
|  | 		Do(ctx). | ||||||
|  | 		Into(result) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Update takes the representation of a fabricEventStream and updates it. Returns the server's representation of the fabricEventStream, and an error, if there is any.
 | ||||||
|  | func (c *fabricEventStreams) Update(ctx context.Context, fabricEventStream *v1.FabricEventStream, opts metav1.UpdateOptions) (result *v1.FabricEventStream, err error) { | ||||||
|  | 	result = &v1.FabricEventStream{} | ||||||
|  | 	err = c.client.Put(). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		Name(fabricEventStream.Name). | ||||||
|  | 		VersionedParams(&opts, scheme.ParameterCodec). | ||||||
|  | 		Body(fabricEventStream). | ||||||
|  | 		Do(ctx). | ||||||
|  | 		Into(result) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Delete takes name of the fabricEventStream and deletes it. Returns an error if one occurs.
 | ||||||
|  | func (c *fabricEventStreams) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { | ||||||
|  | 	return c.client.Delete(). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		Name(name). | ||||||
|  | 		Body(&opts). | ||||||
|  | 		Do(ctx). | ||||||
|  | 		Error() | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeleteCollection deletes a collection of objects.
 | ||||||
|  | func (c *fabricEventStreams) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { | ||||||
|  | 	var timeout time.Duration | ||||||
|  | 	if listOpts.TimeoutSeconds != nil { | ||||||
|  | 		timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second | ||||||
|  | 	} | ||||||
|  | 	return c.client.Delete(). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		VersionedParams(&listOpts, scheme.ParameterCodec). | ||||||
|  | 		Timeout(timeout). | ||||||
|  | 		Body(&opts). | ||||||
|  | 		Do(ctx). | ||||||
|  | 		Error() | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Patch applies the patch and returns the patched fabricEventStream.
 | ||||||
|  | func (c *fabricEventStreams) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.FabricEventStream, err error) { | ||||||
|  | 	result = &v1.FabricEventStream{} | ||||||
|  | 	err = c.client.Patch(pt). | ||||||
|  | 		Namespace(c.ns). | ||||||
|  | 		Resource("fabriceventstreams"). | ||||||
|  | 		Name(name). | ||||||
|  | 		SubResource(subresources...). | ||||||
|  | 		VersionedParams(&opts, scheme.ParameterCodec). | ||||||
|  | 		Body(data). | ||||||
|  | 		Do(ctx). | ||||||
|  | 		Into(result) | ||||||
|  | 	return | ||||||
|  | } | ||||||
|  | @ -0,0 +1,26 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by client-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | // Package fake has the automatically generated clients.
 | ||||||
|  | package fake | ||||||
|  | @ -0,0 +1,136 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by client-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package fake | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	"context" | ||||||
|  | 
 | ||||||
|  | 	zalandoorgv1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
|  | 	v1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
|  | 	labels "k8s.io/apimachinery/pkg/labels" | ||||||
|  | 	schema "k8s.io/apimachinery/pkg/runtime/schema" | ||||||
|  | 	types "k8s.io/apimachinery/pkg/types" | ||||||
|  | 	watch "k8s.io/apimachinery/pkg/watch" | ||||||
|  | 	testing "k8s.io/client-go/testing" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // FakeFabricEventStreams implements FabricEventStreamInterface
 | ||||||
|  | type FakeFabricEventStreams struct { | ||||||
|  | 	Fake *FakeZalandoV1 | ||||||
|  | 	ns   string | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | var fabriceventstreamsResource = schema.GroupVersionResource{Group: "zalando.org", Version: "v1", Resource: "fabriceventstreams"} | ||||||
|  | 
 | ||||||
|  | var fabriceventstreamsKind = schema.GroupVersionKind{Group: "zalando.org", Version: "v1", Kind: "FabricEventStream"} | ||||||
|  | 
 | ||||||
|  | // Get takes name of the fabricEventStream, and returns the corresponding fabricEventStream object, and an error if there is any.
 | ||||||
|  | func (c *FakeFabricEventStreams) Get(ctx context.Context, name string, options v1.GetOptions) (result *zalandoorgv1.FabricEventStream, err error) { | ||||||
|  | 	obj, err := c.Fake. | ||||||
|  | 		Invokes(testing.NewGetAction(fabriceventstreamsResource, c.ns, name), &zalandoorgv1.FabricEventStream{}) | ||||||
|  | 
 | ||||||
|  | 	if obj == nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 	return obj.(*zalandoorgv1.FabricEventStream), err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // List takes label and field selectors, and returns the list of FabricEventStreams that match those selectors.
 | ||||||
|  | func (c *FakeFabricEventStreams) List(ctx context.Context, opts v1.ListOptions) (result *zalandoorgv1.FabricEventStreamList, err error) { | ||||||
|  | 	obj, err := c.Fake. | ||||||
|  | 		Invokes(testing.NewListAction(fabriceventstreamsResource, fabriceventstreamsKind, c.ns, opts), &zalandoorgv1.FabricEventStreamList{}) | ||||||
|  | 
 | ||||||
|  | 	if obj == nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	label, _, _ := testing.ExtractFromListOptions(opts) | ||||||
|  | 	if label == nil { | ||||||
|  | 		label = labels.Everything() | ||||||
|  | 	} | ||||||
|  | 	list := &zalandoorgv1.FabricEventStreamList{ListMeta: obj.(*zalandoorgv1.FabricEventStreamList).ListMeta} | ||||||
|  | 	for _, item := range obj.(*zalandoorgv1.FabricEventStreamList).Items { | ||||||
|  | 		if label.Matches(labels.Set(item.Labels)) { | ||||||
|  | 			list.Items = append(list.Items, item) | ||||||
|  | 		} | ||||||
|  | 	} | ||||||
|  | 	return list, err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Watch returns a watch.Interface that watches the requested fabricEventStreams.
 | ||||||
|  | func (c *FakeFabricEventStreams) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) { | ||||||
|  | 	return c.Fake. | ||||||
|  | 		InvokesWatch(testing.NewWatchAction(fabriceventstreamsResource, c.ns, opts)) | ||||||
|  | 
 | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Create takes the representation of a fabricEventStream and creates it.  Returns the server's representation of the fabricEventStream, and an error, if there is any.
 | ||||||
|  | func (c *FakeFabricEventStreams) Create(ctx context.Context, fabricEventStream *zalandoorgv1.FabricEventStream, opts v1.CreateOptions) (result *zalandoorgv1.FabricEventStream, err error) { | ||||||
|  | 	obj, err := c.Fake. | ||||||
|  | 		Invokes(testing.NewCreateAction(fabriceventstreamsResource, c.ns, fabricEventStream), &zalandoorgv1.FabricEventStream{}) | ||||||
|  | 
 | ||||||
|  | 	if obj == nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 	return obj.(*zalandoorgv1.FabricEventStream), err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Update takes the representation of a fabricEventStream and updates it. Returns the server's representation of the fabricEventStream, and an error, if there is any.
 | ||||||
|  | func (c *FakeFabricEventStreams) Update(ctx context.Context, fabricEventStream *zalandoorgv1.FabricEventStream, opts v1.UpdateOptions) (result *zalandoorgv1.FabricEventStream, err error) { | ||||||
|  | 	obj, err := c.Fake. | ||||||
|  | 		Invokes(testing.NewUpdateAction(fabriceventstreamsResource, c.ns, fabricEventStream), &zalandoorgv1.FabricEventStream{}) | ||||||
|  | 
 | ||||||
|  | 	if obj == nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 	return obj.(*zalandoorgv1.FabricEventStream), err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Delete takes name of the fabricEventStream and deletes it. Returns an error if one occurs.
 | ||||||
|  | func (c *FakeFabricEventStreams) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error { | ||||||
|  | 	_, err := c.Fake. | ||||||
|  | 		Invokes(testing.NewDeleteAction(fabriceventstreamsResource, c.ns, name), &zalandoorgv1.FabricEventStream{}) | ||||||
|  | 
 | ||||||
|  | 	return err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // DeleteCollection deletes a collection of objects.
 | ||||||
|  | func (c *FakeFabricEventStreams) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error { | ||||||
|  | 	action := testing.NewDeleteCollectionAction(fabriceventstreamsResource, c.ns, listOpts) | ||||||
|  | 
 | ||||||
|  | 	_, err := c.Fake.Invokes(action, &zalandoorgv1.FabricEventStreamList{}) | ||||||
|  | 	return err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Patch applies the patch and returns the patched fabricEventStream.
 | ||||||
|  | func (c *FakeFabricEventStreams) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *zalandoorgv1.FabricEventStream, err error) { | ||||||
|  | 	obj, err := c.Fake. | ||||||
|  | 		Invokes(testing.NewPatchSubresourceAction(fabriceventstreamsResource, c.ns, name, pt, data, subresources...), &zalandoorgv1.FabricEventStream{}) | ||||||
|  | 
 | ||||||
|  | 	if obj == nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 	return obj.(*zalandoorgv1.FabricEventStream), err | ||||||
|  | } | ||||||
|  | @ -0,0 +1,46 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by client-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package fake | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	v1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/zalando.org/v1" | ||||||
|  | 	rest "k8s.io/client-go/rest" | ||||||
|  | 	testing "k8s.io/client-go/testing" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | type FakeZalandoV1 struct { | ||||||
|  | 	*testing.Fake | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *FakeZalandoV1) FabricEventStreams(namespace string) v1.FabricEventStreamInterface { | ||||||
|  | 	return &FakeFabricEventStreams{c, namespace} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // RESTClient returns a RESTClient that is used to communicate
 | ||||||
|  | // with API server by this client implementation.
 | ||||||
|  | func (c *FakeZalandoV1) RESTClient() rest.Interface { | ||||||
|  | 	var ret *rest.RESTClient | ||||||
|  | 	return ret | ||||||
|  | } | ||||||
|  | @ -0,0 +1,27 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by client-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | type FabricEventStreamExpansion interface{} | ||||||
|  | @ -0,0 +1,95 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by client-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	v1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
|  | 	"github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/scheme" | ||||||
|  | 	rest "k8s.io/client-go/rest" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | type ZalandoV1Interface interface { | ||||||
|  | 	RESTClient() rest.Interface | ||||||
|  | 	FabricEventStreamsGetter | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // ZalandoV1Client is used to interact with features provided by the zalando.org group.
 | ||||||
|  | type ZalandoV1Client struct { | ||||||
|  | 	restClient rest.Interface | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (c *ZalandoV1Client) FabricEventStreams(namespace string) FabricEventStreamInterface { | ||||||
|  | 	return newFabricEventStreams(c, namespace) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // NewForConfig creates a new ZalandoV1Client for the given config.
 | ||||||
|  | func NewForConfig(c *rest.Config) (*ZalandoV1Client, error) { | ||||||
|  | 	config := *c | ||||||
|  | 	if err := setConfigDefaults(&config); err != nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 	client, err := rest.RESTClientFor(&config) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 	return &ZalandoV1Client{client}, nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // NewForConfigOrDie creates a new ZalandoV1Client for the given config and
 | ||||||
|  | // panics if there is an error in the config.
 | ||||||
|  | func NewForConfigOrDie(c *rest.Config) *ZalandoV1Client { | ||||||
|  | 	client, err := NewForConfig(c) | ||||||
|  | 	if err != nil { | ||||||
|  | 		panic(err) | ||||||
|  | 	} | ||||||
|  | 	return client | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // New creates a new ZalandoV1Client for the given RESTClient.
 | ||||||
|  | func New(c rest.Interface) *ZalandoV1Client { | ||||||
|  | 	return &ZalandoV1Client{c} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func setConfigDefaults(config *rest.Config) error { | ||||||
|  | 	gv := v1.SchemeGroupVersion | ||||||
|  | 	config.GroupVersion = &gv | ||||||
|  | 	config.APIPath = "/apis" | ||||||
|  | 	config.NegotiatedSerializer = scheme.Codecs.WithoutConversion() | ||||||
|  | 
 | ||||||
|  | 	if config.UserAgent == "" { | ||||||
|  | 		config.UserAgent = rest.DefaultKubernetesUserAgent() | ||||||
|  | 	} | ||||||
|  | 
 | ||||||
|  | 	return nil | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // RESTClient returns a RESTClient that is used to communicate
 | ||||||
|  | // with API server by this client implementation.
 | ||||||
|  | func (c *ZalandoV1Client) RESTClient() rest.Interface { | ||||||
|  | 	if c == nil { | ||||||
|  | 		return nil | ||||||
|  | 	} | ||||||
|  | 	return c.restClient | ||||||
|  | } | ||||||
|  | @ -32,6 +32,7 @@ import ( | ||||||
| 	versioned "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned" | 	versioned "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned" | ||||||
| 	acidzalando "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/acid.zalan.do" | 	acidzalando "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/acid.zalan.do" | ||||||
| 	internalinterfaces "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/internalinterfaces" | 	internalinterfaces "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/internalinterfaces" | ||||||
|  | 	zalandoorg "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/zalando.org" | ||||||
| 	v1 "k8s.io/apimachinery/pkg/apis/meta/v1" | 	v1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
| 	runtime "k8s.io/apimachinery/pkg/runtime" | 	runtime "k8s.io/apimachinery/pkg/runtime" | ||||||
| 	schema "k8s.io/apimachinery/pkg/runtime/schema" | 	schema "k8s.io/apimachinery/pkg/runtime/schema" | ||||||
|  | @ -179,8 +180,13 @@ type SharedInformerFactory interface { | ||||||
| 	WaitForCacheSync(stopCh <-chan struct{}) map[reflect.Type]bool | 	WaitForCacheSync(stopCh <-chan struct{}) map[reflect.Type]bool | ||||||
| 
 | 
 | ||||||
| 	Acid() acidzalando.Interface | 	Acid() acidzalando.Interface | ||||||
|  | 	Zalando() zalandoorg.Interface | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| func (f *sharedInformerFactory) Acid() acidzalando.Interface { | func (f *sharedInformerFactory) Acid() acidzalando.Interface { | ||||||
| 	return acidzalando.New(f, f.namespace, f.tweakListOptions) | 	return acidzalando.New(f, f.namespace, f.tweakListOptions) | ||||||
| } | } | ||||||
|  | 
 | ||||||
|  | func (f *sharedInformerFactory) Zalando() zalandoorg.Interface { | ||||||
|  | 	return zalandoorg.New(f, f.namespace, f.tweakListOptions) | ||||||
|  | } | ||||||
|  |  | ||||||
|  | @ -28,6 +28,7 @@ import ( | ||||||
| 	"fmt" | 	"fmt" | ||||||
| 
 | 
 | ||||||
| 	v1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | 	v1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | ||||||
|  | 	zalandoorgv1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
| 	schema "k8s.io/apimachinery/pkg/runtime/schema" | 	schema "k8s.io/apimachinery/pkg/runtime/schema" | ||||||
| 	cache "k8s.io/client-go/tools/cache" | 	cache "k8s.io/client-go/tools/cache" | ||||||
| ) | ) | ||||||
|  | @ -64,6 +65,10 @@ func (f *sharedInformerFactory) ForResource(resource schema.GroupVersionResource | ||||||
| 	case v1.SchemeGroupVersion.WithResource("postgresqls"): | 	case v1.SchemeGroupVersion.WithResource("postgresqls"): | ||||||
| 		return &genericInformer{resource: resource.GroupResource(), informer: f.Acid().V1().Postgresqls().Informer()}, nil | 		return &genericInformer{resource: resource.GroupResource(), informer: f.Acid().V1().Postgresqls().Informer()}, nil | ||||||
| 
 | 
 | ||||||
|  | 		// Group=zalando.org, Version=v1
 | ||||||
|  | 	case zalandoorgv1.SchemeGroupVersion.WithResource("fabriceventstreams"): | ||||||
|  | 		return &genericInformer{resource: resource.GroupResource(), informer: f.Zalando().V1().FabricEventStreams().Informer()}, nil | ||||||
|  | 
 | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	return nil, fmt.Errorf("no informer found for %v", resource) | 	return nil, fmt.Errorf("no informer found for %v", resource) | ||||||
|  |  | ||||||
|  | @ -0,0 +1,52 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by informer-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package zalando | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	internalinterfaces "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/internalinterfaces" | ||||||
|  | 	v1 "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/zalando.org/v1" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // Interface provides access to each of this group's versions.
 | ||||||
|  | type Interface interface { | ||||||
|  | 	// V1 provides access to shared informers for resources in V1.
 | ||||||
|  | 	V1() v1.Interface | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | type group struct { | ||||||
|  | 	factory          internalinterfaces.SharedInformerFactory | ||||||
|  | 	namespace        string | ||||||
|  | 	tweakListOptions internalinterfaces.TweakListOptionsFunc | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // New returns a new Interface.
 | ||||||
|  | func New(f internalinterfaces.SharedInformerFactory, namespace string, tweakListOptions internalinterfaces.TweakListOptionsFunc) Interface { | ||||||
|  | 	return &group{factory: f, namespace: namespace, tweakListOptions: tweakListOptions} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // V1 returns a new v1.Interface.
 | ||||||
|  | func (g *group) V1() v1.Interface { | ||||||
|  | 	return v1.New(g.factory, g.namespace, g.tweakListOptions) | ||||||
|  | } | ||||||
|  | @ -0,0 +1,96 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by informer-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	"context" | ||||||
|  | 	time "time" | ||||||
|  | 
 | ||||||
|  | 	zalandoorgv1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
|  | 	versioned "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned" | ||||||
|  | 	internalinterfaces "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/internalinterfaces" | ||||||
|  | 	v1 "github.com/zalando/postgres-operator/pkg/generated/listers/zalando.org/v1" | ||||||
|  | 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||||
|  | 	runtime "k8s.io/apimachinery/pkg/runtime" | ||||||
|  | 	watch "k8s.io/apimachinery/pkg/watch" | ||||||
|  | 	cache "k8s.io/client-go/tools/cache" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamInformer provides access to a shared informer and lister for
 | ||||||
|  | // FabricEventStreams.
 | ||||||
|  | type FabricEventStreamInformer interface { | ||||||
|  | 	Informer() cache.SharedIndexInformer | ||||||
|  | 	Lister() v1.FabricEventStreamLister | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | type fabricEventStreamInformer struct { | ||||||
|  | 	factory          internalinterfaces.SharedInformerFactory | ||||||
|  | 	tweakListOptions internalinterfaces.TweakListOptionsFunc | ||||||
|  | 	namespace        string | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // NewFabricEventStreamInformer constructs a new informer for FabricEventStream type.
 | ||||||
|  | // Always prefer using an informer factory to get a shared informer instead of getting an independent
 | ||||||
|  | // one. This reduces memory footprint and number of connections to the server.
 | ||||||
|  | func NewFabricEventStreamInformer(client versioned.Interface, namespace string, resyncPeriod time.Duration, indexers cache.Indexers) cache.SharedIndexInformer { | ||||||
|  | 	return NewFilteredFabricEventStreamInformer(client, namespace, resyncPeriod, indexers, nil) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // NewFilteredFabricEventStreamInformer constructs a new informer for FabricEventStream type.
 | ||||||
|  | // Always prefer using an informer factory to get a shared informer instead of getting an independent
 | ||||||
|  | // one. This reduces memory footprint and number of connections to the server.
 | ||||||
|  | func NewFilteredFabricEventStreamInformer(client versioned.Interface, namespace string, resyncPeriod time.Duration, indexers cache.Indexers, tweakListOptions internalinterfaces.TweakListOptionsFunc) cache.SharedIndexInformer { | ||||||
|  | 	return cache.NewSharedIndexInformer( | ||||||
|  | 		&cache.ListWatch{ | ||||||
|  | 			ListFunc: func(options metav1.ListOptions) (runtime.Object, error) { | ||||||
|  | 				if tweakListOptions != nil { | ||||||
|  | 					tweakListOptions(&options) | ||||||
|  | 				} | ||||||
|  | 				return client.ZalandoV1().FabricEventStreams(namespace).List(context.TODO(), options) | ||||||
|  | 			}, | ||||||
|  | 			WatchFunc: func(options metav1.ListOptions) (watch.Interface, error) { | ||||||
|  | 				if tweakListOptions != nil { | ||||||
|  | 					tweakListOptions(&options) | ||||||
|  | 				} | ||||||
|  | 				return client.ZalandoV1().FabricEventStreams(namespace).Watch(context.TODO(), options) | ||||||
|  | 			}, | ||||||
|  | 		}, | ||||||
|  | 		&zalandoorgv1.FabricEventStream{}, | ||||||
|  | 		resyncPeriod, | ||||||
|  | 		indexers, | ||||||
|  | 	) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (f *fabricEventStreamInformer) defaultInformer(client versioned.Interface, resyncPeriod time.Duration) cache.SharedIndexInformer { | ||||||
|  | 	return NewFilteredFabricEventStreamInformer(client, f.namespace, resyncPeriod, cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc}, f.tweakListOptions) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (f *fabricEventStreamInformer) Informer() cache.SharedIndexInformer { | ||||||
|  | 	return f.factory.InformerFor(&zalandoorgv1.FabricEventStream{}, f.defaultInformer) | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | func (f *fabricEventStreamInformer) Lister() v1.FabricEventStreamLister { | ||||||
|  | 	return v1.NewFabricEventStreamLister(f.Informer().GetIndexer()) | ||||||
|  | } | ||||||
|  | @ -0,0 +1,51 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by informer-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	internalinterfaces "github.com/zalando/postgres-operator/pkg/generated/informers/externalversions/internalinterfaces" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // Interface provides access to all the informers in this group version.
 | ||||||
|  | type Interface interface { | ||||||
|  | 	// FabricEventStreams returns a FabricEventStreamInformer.
 | ||||||
|  | 	FabricEventStreams() FabricEventStreamInformer | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | type version struct { | ||||||
|  | 	factory          internalinterfaces.SharedInformerFactory | ||||||
|  | 	namespace        string | ||||||
|  | 	tweakListOptions internalinterfaces.TweakListOptionsFunc | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // New returns a new Interface.
 | ||||||
|  | func New(f internalinterfaces.SharedInformerFactory, namespace string, tweakListOptions internalinterfaces.TweakListOptionsFunc) Interface { | ||||||
|  | 	return &version{factory: f, namespace: namespace, tweakListOptions: tweakListOptions} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // FabricEventStreams returns a FabricEventStreamInformer.
 | ||||||
|  | func (v *version) FabricEventStreams() FabricEventStreamInformer { | ||||||
|  | 	return &fabricEventStreamInformer{factory: v.factory, namespace: v.namespace, tweakListOptions: v.tweakListOptions} | ||||||
|  | } | ||||||
|  | @ -0,0 +1,33 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by lister-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamListerExpansion allows custom methods to be added to
 | ||||||
|  | // FabricEventStreamLister.
 | ||||||
|  | type FabricEventStreamListerExpansion interface{} | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamNamespaceListerExpansion allows custom methods to be added to
 | ||||||
|  | // FabricEventStreamNamespaceLister.
 | ||||||
|  | type FabricEventStreamNamespaceListerExpansion interface{} | ||||||
|  | @ -0,0 +1,105 @@ | ||||||
|  | /* | ||||||
|  | Copyright 2022 Compose, Zalando SE | ||||||
|  | 
 | ||||||
|  | Permission is hereby granted, free of charge, to any person obtaining a copy | ||||||
|  | of this software and associated documentation files (the "Software"), to deal | ||||||
|  | in the Software without restriction, including without limitation the rights | ||||||
|  | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||||||
|  | copies of the Software, and to permit persons to whom the Software is | ||||||
|  | furnished to do so, subject to the following conditions: | ||||||
|  | 
 | ||||||
|  | The above copyright notice and this permission notice shall be included in all | ||||||
|  | copies or substantial portions of the Software. | ||||||
|  | 
 | ||||||
|  | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||||||
|  | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||||||
|  | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||||||
|  | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||||||
|  | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||||||
|  | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||||||
|  | SOFTWARE. | ||||||
|  | */ | ||||||
|  | 
 | ||||||
|  | // Code generated by lister-gen. DO NOT EDIT.
 | ||||||
|  | 
 | ||||||
|  | package v1 | ||||||
|  | 
 | ||||||
|  | import ( | ||||||
|  | 	v1 "github.com/zalando/postgres-operator/pkg/apis/zalando.org/v1" | ||||||
|  | 	"k8s.io/apimachinery/pkg/api/errors" | ||||||
|  | 	"k8s.io/apimachinery/pkg/labels" | ||||||
|  | 	"k8s.io/client-go/tools/cache" | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamLister helps list FabricEventStreams.
 | ||||||
|  | // All objects returned here must be treated as read-only.
 | ||||||
|  | type FabricEventStreamLister interface { | ||||||
|  | 	// List lists all FabricEventStreams in the indexer.
 | ||||||
|  | 	// Objects returned here must be treated as read-only.
 | ||||||
|  | 	List(selector labels.Selector) (ret []*v1.FabricEventStream, err error) | ||||||
|  | 	// FabricEventStreams returns an object that can list and get FabricEventStreams.
 | ||||||
|  | 	FabricEventStreams(namespace string) FabricEventStreamNamespaceLister | ||||||
|  | 	FabricEventStreamListerExpansion | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // fabricEventStreamLister implements the FabricEventStreamLister interface.
 | ||||||
|  | type fabricEventStreamLister struct { | ||||||
|  | 	indexer cache.Indexer | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // NewFabricEventStreamLister returns a new FabricEventStreamLister.
 | ||||||
|  | func NewFabricEventStreamLister(indexer cache.Indexer) FabricEventStreamLister { | ||||||
|  | 	return &fabricEventStreamLister{indexer: indexer} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // List lists all FabricEventStreams in the indexer.
 | ||||||
|  | func (s *fabricEventStreamLister) List(selector labels.Selector) (ret []*v1.FabricEventStream, err error) { | ||||||
|  | 	err = cache.ListAll(s.indexer, selector, func(m interface{}) { | ||||||
|  | 		ret = append(ret, m.(*v1.FabricEventStream)) | ||||||
|  | 	}) | ||||||
|  | 	return ret, err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // FabricEventStreams returns an object that can list and get FabricEventStreams.
 | ||||||
|  | func (s *fabricEventStreamLister) FabricEventStreams(namespace string) FabricEventStreamNamespaceLister { | ||||||
|  | 	return fabricEventStreamNamespaceLister{indexer: s.indexer, namespace: namespace} | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // FabricEventStreamNamespaceLister helps list and get FabricEventStreams.
 | ||||||
|  | // All objects returned here must be treated as read-only.
 | ||||||
|  | type FabricEventStreamNamespaceLister interface { | ||||||
|  | 	// List lists all FabricEventStreams in the indexer for a given namespace.
 | ||||||
|  | 	// Objects returned here must be treated as read-only.
 | ||||||
|  | 	List(selector labels.Selector) (ret []*v1.FabricEventStream, err error) | ||||||
|  | 	// Get retrieves the FabricEventStream from the indexer for a given namespace and name.
 | ||||||
|  | 	// Objects returned here must be treated as read-only.
 | ||||||
|  | 	Get(name string) (*v1.FabricEventStream, error) | ||||||
|  | 	FabricEventStreamNamespaceListerExpansion | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // fabricEventStreamNamespaceLister implements the FabricEventStreamNamespaceLister
 | ||||||
|  | // interface.
 | ||||||
|  | type fabricEventStreamNamespaceLister struct { | ||||||
|  | 	indexer   cache.Indexer | ||||||
|  | 	namespace string | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // List lists all FabricEventStreams in the indexer for a given namespace.
 | ||||||
|  | func (s fabricEventStreamNamespaceLister) List(selector labels.Selector) (ret []*v1.FabricEventStream, err error) { | ||||||
|  | 	err = cache.ListAllByNamespace(s.indexer, s.namespace, selector, func(m interface{}) { | ||||||
|  | 		ret = append(ret, m.(*v1.FabricEventStream)) | ||||||
|  | 	}) | ||||||
|  | 	return ret, err | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | // Get retrieves the FabricEventStream from the indexer for a given namespace and name.
 | ||||||
|  | func (s fabricEventStreamNamespaceLister) Get(name string) (*v1.FabricEventStream, error) { | ||||||
|  | 	obj, exists, err := s.indexer.GetByKey(s.namespace + "/" + name) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return nil, err | ||||||
|  | 	} | ||||||
|  | 	if !exists { | ||||||
|  | 		return nil, errors.NewNotFound(v1.Resource("fabriceventstream"), name) | ||||||
|  | 	} | ||||||
|  | 	return obj.(*v1.FabricEventStream), nil | ||||||
|  | } | ||||||
|  | @ -0,0 +1,14 @@ | ||||||
|  | package constants | ||||||
|  | 
 | ||||||
|  | // PostgreSQL specific constants
 | ||||||
|  | const ( | ||||||
|  | 	EventStreamCRDApiVersion     = "zalando.org/v1" | ||||||
|  | 	EventStreamCRDKind           = "FabricEventStream" | ||||||
|  | 	EventStreamCRDName           = "fabriceventstreams.zalando.org" | ||||||
|  | 	EventStreamSourcePGType      = "PostgresLogicalReplication" | ||||||
|  | 	EventStreamSourceSlotPrefix  = "fes" | ||||||
|  | 	EventStreamSourcePluginType  = "pgoutput" | ||||||
|  | 	EventStreamSourceAuthType    = "DatabaseAuthenticationSecret" | ||||||
|  | 	EventStreamFlowPgGenericType = "PostgresWalToGenericNakadiEvent" | ||||||
|  | 	EventStreamSinkNakadiType    = "Nakadi" | ||||||
|  | ) | ||||||
|  | @ -12,8 +12,9 @@ import ( | ||||||
| 	clientbatchv1beta1 "k8s.io/client-go/kubernetes/typed/batch/v1beta1" | 	clientbatchv1beta1 "k8s.io/client-go/kubernetes/typed/batch/v1beta1" | ||||||
| 
 | 
 | ||||||
| 	apiacidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | 	apiacidv1 "github.com/zalando/postgres-operator/pkg/apis/acid.zalan.do/v1" | ||||||
| 	acidv1client "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned" | 	zalandoclient "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned" | ||||||
| 	acidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1" | 	acidv1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/acid.zalan.do/v1" | ||||||
|  | 	zalandov1 "github.com/zalando/postgres-operator/pkg/generated/clientset/versioned/typed/zalando.org/v1" | ||||||
| 	"github.com/zalando/postgres-operator/pkg/spec" | 	"github.com/zalando/postgres-operator/pkg/spec" | ||||||
| 	apiappsv1 "k8s.io/api/apps/v1" | 	apiappsv1 "k8s.io/api/apps/v1" | ||||||
| 	v1 "k8s.io/api/core/v1" | 	v1 "k8s.io/api/core/v1" | ||||||
|  | @ -58,9 +59,11 @@ type KubernetesClient struct { | ||||||
| 	acidv1.OperatorConfigurationsGetter | 	acidv1.OperatorConfigurationsGetter | ||||||
| 	acidv1.PostgresTeamsGetter | 	acidv1.PostgresTeamsGetter | ||||||
| 	acidv1.PostgresqlsGetter | 	acidv1.PostgresqlsGetter | ||||||
|  | 	zalandov1.FabricEventStreamsGetter | ||||||
| 
 | 
 | ||||||
| 	RESTClient      rest.Interface | 	RESTClient         rest.Interface | ||||||
| 	AcidV1ClientSet *acidv1client.Clientset | 	AcidV1ClientSet    *zalandoclient.Clientset | ||||||
|  | 	Zalandov1ClientSet *zalandoclient.Clientset | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| type mockSecret struct { | type mockSecret struct { | ||||||
|  | @ -158,14 +161,19 @@ func NewFromConfig(cfg *rest.Config) (KubernetesClient, error) { | ||||||
| 
 | 
 | ||||||
| 	kubeClient.CustomResourceDefinitionsGetter = apiextClient.ApiextensionsV1() | 	kubeClient.CustomResourceDefinitionsGetter = apiextClient.ApiextensionsV1() | ||||||
| 
 | 
 | ||||||
| 	kubeClient.AcidV1ClientSet = acidv1client.NewForConfigOrDie(cfg) | 	kubeClient.AcidV1ClientSet = zalandoclient.NewForConfigOrDie(cfg) | ||||||
| 	if err != nil { | 	if err != nil { | ||||||
| 		return kubeClient, fmt.Errorf("could not create acid.zalan.do clientset: %v", err) | 		return kubeClient, fmt.Errorf("could not create acid.zalan.do clientset: %v", err) | ||||||
| 	} | 	} | ||||||
|  | 	kubeClient.Zalandov1ClientSet = zalandoclient.NewForConfigOrDie(cfg) | ||||||
|  | 	if err != nil { | ||||||
|  | 		return kubeClient, fmt.Errorf("could not create zalando.org clientset: %v", err) | ||||||
|  | 	} | ||||||
| 
 | 
 | ||||||
| 	kubeClient.OperatorConfigurationsGetter = kubeClient.AcidV1ClientSet.AcidV1() | 	kubeClient.OperatorConfigurationsGetter = kubeClient.AcidV1ClientSet.AcidV1() | ||||||
| 	kubeClient.PostgresTeamsGetter = kubeClient.AcidV1ClientSet.AcidV1() | 	kubeClient.PostgresTeamsGetter = kubeClient.AcidV1ClientSet.AcidV1() | ||||||
| 	kubeClient.PostgresqlsGetter = kubeClient.AcidV1ClientSet.AcidV1() | 	kubeClient.PostgresqlsGetter = kubeClient.AcidV1ClientSet.AcidV1() | ||||||
|  | 	kubeClient.FabricEventStreamsGetter = kubeClient.Zalandov1ClientSet.ZalandoV1() | ||||||
| 
 | 
 | ||||||
| 	return kubeClient, nil | 	return kubeClient, nil | ||||||
| } | } | ||||||
|  |  | ||||||
|  | @ -185,7 +185,7 @@ func TestGetConfig(t *testing.T) { | ||||||
| 		Slots: map[string]map[string]string{ | 		Slots: map[string]map[string]string{ | ||||||
| 			"cdc": { | 			"cdc": { | ||||||
| 				"database": "foo", | 				"database": "foo", | ||||||
| 				"plugin":   "wal2json", | 				"plugin":   "pgoutput", | ||||||
| 				"type":     "logical", | 				"type":     "logical", | ||||||
| 			}, | 			}, | ||||||
| 		}, | 		}, | ||||||
|  | @ -218,7 +218,7 @@ func TestGetConfig(t *testing.T) { | ||||||
| 		"wal_log_hints":                   "on", | 		"wal_log_hints":                   "on", | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	configJson := `{"loop_wait": 10, "maximum_lag_on_failover": 33554432, "postgresql": {"parameters": {"archive_mode": "on", "archive_timeout": "1800s", "autovacuum_analyze_scale_factor": 0.02, "autovacuum_max_workers": 5, "autovacuum_vacuum_scale_factor": 0.05, "checkpoint_completion_target": 0.9, "hot_standby": "on", "log_autovacuum_min_duration": 0, "log_checkpoints": "on", "log_connections": "on", "log_disconnections": "on", "log_line_prefix": "%t [%p]: [%l-1] %c %x %d %u %a %h ", "log_lock_waits": "on", "log_min_duration_statement": 500, "log_statement": "ddl", "log_temp_files": 0, "max_connections": 100, "max_replication_slots": 10, "max_wal_senders": 10, "tcp_keepalives_idle": 900, "tcp_keepalives_interval": 100, "track_functions": "all", "wal_level": "hot_standby", "wal_log_hints": "on"}, "use_pg_rewind": true, "use_slots": true}, "retry_timeout": 10, "slots": {"cdc": {"database": "foo", "plugin": "wal2json", "type": "logical"}}, "ttl": 30}` | 	configJson := `{"loop_wait": 10, "maximum_lag_on_failover": 33554432, "postgresql": {"parameters": {"archive_mode": "on", "archive_timeout": "1800s", "autovacuum_analyze_scale_factor": 0.02, "autovacuum_max_workers": 5, "autovacuum_vacuum_scale_factor": 0.05, "checkpoint_completion_target": 0.9, "hot_standby": "on", "log_autovacuum_min_duration": 0, "log_checkpoints": "on", "log_connections": "on", "log_disconnections": "on", "log_line_prefix": "%t [%p]: [%l-1] %c %x %d %u %a %h ", "log_lock_waits": "on", "log_min_duration_statement": 500, "log_statement": "ddl", "log_temp_files": 0, "max_connections": 100, "max_replication_slots": 10, "max_wal_senders": 10, "tcp_keepalives_idle": 900, "tcp_keepalives_interval": 100, "track_functions": "all", "wal_level": "hot_standby", "wal_log_hints": "on"}, "use_pg_rewind": true, "use_slots": true}, "retry_timeout": 10, "slots": {"cdc": {"database": "foo", "plugin": "pgoutput", "type": "logical"}}, "ttl": 30}` | ||||||
| 	r := ioutil.NopCloser(bytes.NewReader([]byte(configJson))) | 	r := ioutil.NopCloser(bytes.NewReader([]byte(configJson))) | ||||||
| 
 | 
 | ||||||
| 	response := http.Response{ | 	response := http.Response{ | ||||||
|  | @ -253,7 +253,7 @@ func TestSetPostgresParameters(t *testing.T) { | ||||||
| 		"wal_level":       "logical", | 		"wal_level":       "logical", | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	configJson := `{"loop_wait": 10, "maximum_lag_on_failover": 33554432, "postgresql": {"parameters": {"archive_mode": "on", "archive_timeout": "1800s", "autovacuum_analyze_scale_factor": 0.02, "autovacuum_max_workers": 5, "autovacuum_vacuum_scale_factor": 0.05, "checkpoint_completion_target": 0.9, "hot_standby": "on", "log_autovacuum_min_duration": 0, "log_checkpoints": "on", "log_connections": "on", "log_disconnections": "on", "log_line_prefix": "%t [%p]: [%l-1] %c %x %d %u %a %h ", "log_lock_waits": "on", "log_min_duration_statement": 500, "log_statement": "ddl", "log_temp_files": 0, "max_connections": 50, "max_replication_slots": 10, "max_wal_senders": 10, "tcp_keepalives_idle": 900, "tcp_keepalives_interval": 100, "track_functions": "all", "wal_level": "logical", "wal_log_hints": "on"}, "use_pg_rewind": true, "use_slots": true}, "retry_timeout": 10, "slots": {"cdc": {"database": "foo", "plugin": "wal2json", "type": "logical"}}, "ttl": 30}` | 	configJson := `{"loop_wait": 10, "maximum_lag_on_failover": 33554432, "postgresql": {"parameters": {"archive_mode": "on", "archive_timeout": "1800s", "autovacuum_analyze_scale_factor": 0.02, "autovacuum_max_workers": 5, "autovacuum_vacuum_scale_factor": 0.05, "checkpoint_completion_target": 0.9, "hot_standby": "on", "log_autovacuum_min_duration": 0, "log_checkpoints": "on", "log_connections": "on", "log_disconnections": "on", "log_line_prefix": "%t [%p]: [%l-1] %c %x %d %u %a %h ", "log_lock_waits": "on", "log_min_duration_statement": 500, "log_statement": "ddl", "log_temp_files": 0, "max_connections": 50, "max_replication_slots": 10, "max_wal_senders": 10, "tcp_keepalives_idle": 900, "tcp_keepalives_interval": 100, "track_functions": "all", "wal_level": "logical", "wal_log_hints": "on"}, "use_pg_rewind": true, "use_slots": true}, "retry_timeout": 10, "slots": {"cdc": {"database": "foo", "plugin": "pgoutput", "type": "logical"}}, "ttl": 30}` | ||||||
| 	r := ioutil.NopCloser(bytes.NewReader([]byte(configJson))) | 	r := ioutil.NopCloser(bytes.NewReader([]byte(configJson))) | ||||||
| 
 | 
 | ||||||
| 	response := http.Response{ | 	response := http.Response{ | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue