跳转至

PostgreSQL HA packaged by Bitnami

This PostgreSQL cluster solution includes the PostgreSQL replication manager, an open-source tool for managing replication and failover on PostgreSQL clusters.

Overview of PostgreSQL HA

Trademarks: This software listing is packaged by Bitnami. The respective trademarks mentioned in the offering are owned by the respective companies, and use of them does not imply any affiliation or endorsement.

TL;DR

使用加速地址添加仓库:

helm repo add bitnami "https://helm-charts.itboon.top/bitnami" --force-update
helm update bitnami
helm install my-release bitnami/postgresql-ha

Introduction

This Helm chart installs PostgreSQL with HA architecture in a Kubernetes cluster. Welcome to contribute to Helm Chart for PostgreSQL HA.

This Helm chart has been developed based on bitnami/postgresql chart but including some changes to guarantee high availability such as:

  • A new deployment, service have been added to deploy Pgpool-II to act as proxy for PostgreSQL backend. It helps to reduce connection overhead, acts as a load balancer for PostgreSQL, and ensures database node failover.
  • Replacing bitnami/postgresql with bitnami/postgresql-repmgr which includes and configures repmgr. Repmgr ensures standby nodes assume the primary role when the primary node is unhealthy.

Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters.

Looking to use PostgreSQL HA in production? Try VMware Application Catalog, the enterprise edition of Bitnami Application Catalog.

Prerequisites

  • Kubernetes 1.19+
  • Helm 3.2.0+

Installing the Chart

To install the chart with the release name my-release:

helm install my-release bitnami/postgresql-ha

Uninstalling the Chart

To uninstall/delete the my-release deployment:

helm delete --purge my-release

Additionally, if persistence.resourcePolicy is set to keep, you should manually delete the PVCs.

Parameters

Global parameters

Name Description Value
global.imageRegistry Global Docker image registry ""
global.imagePullSecrets Global Docker registry secret names as an array []
global.storageClass Global StorageClass for Persistent Volume(s) ""
global.postgresql.username PostgreSQL username (overrides postgresql.username) ""
global.postgresql.password PostgreSQL password (overrides postgresql.password) ""
global.postgresql.database PostgreSQL database (overrides postgresql.database) ""
global.postgresql.repmgrUsername PostgreSQL repmgr username (overrides postgresql.repmgrUsername) ""
global.postgresql.repmgrPassword PostgreSQL repmgr password (overrides postgresql.repmgrpassword) ""
global.postgresql.repmgrDatabase PostgreSQL repmgr database (overrides postgresql.repmgrDatabase) ""
global.postgresql.existingSecret Name of existing secret to use for PostgreSQL passwords (overrides postgresql.existingSecret) ""
global.ldap.bindpw LDAP bind password (overrides ldap.bindpw) ""
global.ldap.existingSecret Name of existing secret to use for LDAP passwords (overrides ldap.existingSecret) ""
global.pgpool.adminUsername Pgpool Admin username (overrides pgpool.adminUsername) ""
global.pgpool.adminPassword Pgpool Admin password (overrides pgpool.adminPassword) ""
global.pgpool.existingSecret Pgpool existing secret ""

Common parameters

Name Description Value
kubeVersion Override Kubernetes version ""
nameOverride String to partially override common.names.fullname template (will maintain the release name) ""
fullnameOverride String to fully override common.names.fullname template ""
namespaceOverride String to fully override common.names.namespace ""
commonLabels Common labels to add to all resources (sub-charts are not considered). Evaluated as a template {}
commonAnnotations Common annotations to add to all resources (sub-charts are not considered). Evaluated as a template {}
clusterDomain Kubernetes Cluster Domain cluster.local
extraDeploy Array of extra objects to deploy with the release (evaluated as a template) []
diagnosticMode.enabled Enable diagnostic mode (all probes will be disabled and the command will be overridden) false
diagnosticMode.command Command to override all containers in the deployment []
diagnosticMode.args Args to override all containers in the deployment []

PostgreSQL with Repmgr parameters

Name Description Value
postgresql.image.registry PostgreSQL with Repmgr image registry docker.io
postgresql.image.repository PostgreSQL with Repmgr image repository bitnami/postgresql-repmgr
postgresql.image.tag PostgreSQL with Repmgr image tag 15.3.0-debian-11-r16
postgresql.image.digest PostgreSQL image digest in the way sha256:aa.... Please note this parameter, if set, will override the tag ""
postgresql.image.pullPolicy PostgreSQL with Repmgr image pull policy IfNotPresent
postgresql.image.pullSecrets Specify docker-registry secret names as an array []
postgresql.image.debug Specify if debug logs should be enabled false
postgresql.labels Labels to add to the StatefulSet. Evaluated as template {}
postgresql.podLabels Labels to add to the StatefulSet pods. Evaluated as template {}
postgresql.serviceAnnotations Provide any additional annotations for PostgreSQL service {}
postgresql.replicaCount Number of replicas to deploy. Use an odd number. Having 3 replicas is the minimum to get quorum when promoting a new primary. 3
postgresql.updateStrategy.type Postgresql statefulset strategy type RollingUpdate
postgresql.containerPorts.postgresql PostgreSQL port 5432
postgresql.hostAliases Deployment pod host aliases []
postgresql.hostNetwork Specify if host network should be enabled for PostgreSQL pod false
postgresql.hostIPC Specify if host IPC should be enabled for PostgreSQL pod false
postgresql.podAnnotations Additional pod annotations {}
postgresql.podAffinityPreset PostgreSQL pod affinity preset. Ignored if postgresql.affinity is set. Allowed values: soft or hard ""
postgresql.podAntiAffinityPreset PostgreSQL pod anti-affinity preset. Ignored if postgresql.affinity is set. Allowed values: soft or hard soft
postgresql.nodeAffinityPreset.type PostgreSQL node affinity preset type. Ignored if postgresql.affinity is set. Allowed values: soft or hard ""
postgresql.nodeAffinityPreset.key PostgreSQL node label key to match Ignored if postgresql.affinity is set. ""
postgresql.nodeAffinityPreset.values PostgreSQL node label values to match. Ignored if postgresql.affinity is set. []
postgresql.affinity Affinity for PostgreSQL pods assignment {}
postgresql.nodeSelector Node labels for PostgreSQL pods assignment {}
postgresql.tolerations Tolerations for PostgreSQL pods assignment []
postgresql.topologySpreadConstraints Topology Spread Constraints for pod assignment spread across your cluster among failure-domains. Evaluated as a template []
postgresql.priorityClassName Pod priority class ""
postgresql.schedulerName Use an alternate scheduler, e.g. "stork". ""
postgresql.terminationGracePeriodSeconds Seconds PostgreSQL pod needs to terminate gracefully ""
postgresql.podSecurityContext.enabled Enable security context for PostgreSQL with Repmgr true
postgresql.podSecurityContext.fsGroup Group ID for the PostgreSQL with Repmgr filesystem 1001
postgresql.containerSecurityContext.enabled Enable container security context true
postgresql.containerSecurityContext.runAsUser User ID for the PostgreSQL with Repmgr container 1001
postgresql.containerSecurityContext.runAsNonRoot Set PostgreSQL with Repmgr containers' Security Context runAsNonRoot true
postgresql.containerSecurityContext.readOnlyRootFilesystem Set PostgreSQL with Repmgr containers' Security Context runAsNonRoot false
postgresql.command Override default container command (useful when using custom images) []
postgresql.args Override default container args (useful when using custom images) []
postgresql.lifecycleHooks LifecycleHook to set additional configuration at startup, e.g. LDAP settings via REST API. Evaluated as a template {}
postgresql.extraEnvVars Array containing extra environment variables []
postgresql.extraEnvVarsCM ConfigMap with extra environment variables ""
postgresql.extraEnvVarsSecret Secret with extra environment variables ""
postgresql.extraVolumes Extra volumes to add to the deployment []
postgresql.extraVolumeMounts Extra volume mounts to add to the container. Normally used with extraVolumes. []
postgresql.initContainers Extra init containers to add to the deployment []
postgresql.sidecars Extra sidecar containers to add to the deployment []
postgresql.resources.limits The resources limits for the container {}
postgresql.resources.requests The requested resources for the container {}
postgresql.livenessProbe.enabled Enable livenessProbe true
postgresql.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 30
postgresql.livenessProbe.periodSeconds Period seconds for livenessProbe 10
postgresql.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
postgresql.livenessProbe.failureThreshold Failure threshold for livenessProbe 6
postgresql.livenessProbe.successThreshold Success threshold for livenessProbe 1
postgresql.readinessProbe.enabled Enable readinessProbe true
postgresql.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 5
postgresql.readinessProbe.periodSeconds Period seconds for readinessProbe 10
postgresql.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 5
postgresql.readinessProbe.failureThreshold Failure threshold for readinessProbe 6
postgresql.readinessProbe.successThreshold Success threshold for readinessProbe 1
postgresql.startupProbe.enabled Enable startupProbe false
postgresql.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 5
postgresql.startupProbe.periodSeconds Period seconds for startupProbe 10
postgresql.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
postgresql.startupProbe.failureThreshold Failure threshold for startupProbe 10
postgresql.startupProbe.successThreshold Success threshold for startupProbe 1
postgresql.customLivenessProbe Override default liveness probe {}
postgresql.customReadinessProbe Override default readiness probe {}
postgresql.customStartupProbe Override default startup probe {}
postgresql.pdb.create Specifies whether to create a Pod disruption budget for PostgreSQL with Repmgr false
postgresql.pdb.minAvailable Minimum number / percentage of pods that should remain scheduled 1
postgresql.pdb.maxUnavailable Maximum number / percentage of pods that may be made unavailable ""
postgresql.username PostgreSQL username postgres
postgresql.password PostgreSQL password ""
postgresql.database PostgreSQL database ""
postgresql.existingSecret PostgreSQL password using existing secret ""
postgresql.postgresPassword PostgreSQL password for the postgres user when username is not postgres ""
postgresql.usePasswordFile Set to true to mount PostgreSQL secret as a file instead of passing environment variable ""
postgresql.repmgrUsePassfile Set to true to configure repmgrl to use passfile instead of password vars*:::username:password" and use it to configure Repmgr instead of using password (Requires Postgresql 10+, otherwise ignored) ""
postgresql.repmgrPassfilePath Custom path where passfile will be stored ""
postgresql.upgradeRepmgrExtension Upgrade repmgr extension in the database false
postgresql.pgHbaTrustAll Configures PostgreSQL HBA to trust every user false
postgresql.syncReplication Make the replication synchronous. This will wait until the data is synchronized in all the replicas before other query can be run. This ensures the data availability at the expenses of speed. false
postgresql.syncReplicationMode This specifies the method to choose synchronous standbys from the listed servers. Valid values: empty, FIRST, ANY. ""
postgresql.repmgrUsername PostgreSQL Repmgr username repmgr
postgresql.repmgrPassword PostgreSQL Repmgr password ""
postgresql.repmgrDatabase PostgreSQL Repmgr database repmgr
postgresql.repmgrLogLevel Repmgr log level (DEBUG, INFO, NOTICE, WARNING, ERROR, ALERT, CRIT or EMERG) NOTICE
postgresql.repmgrConnectTimeout Repmgr backend connection timeout (in seconds) 5
postgresql.repmgrReconnectAttempts Repmgr backend reconnection attempts 2
postgresql.repmgrReconnectInterval Repmgr backend reconnection interval (in seconds) 3
postgresql.repmgrFenceOldPrimary Set if fencing of old primary in multiple primary situation is desired false
postgresql.repmgrChildNodesCheckInterval Repmgr child nodes check interval (in seconds) 5
postgresql.repmgrChildNodesConnectedMinCount Repmgr minimum number of connected child nodes before being considered as failed primary for fencing 1
postgresql.repmgrChildNodesDisconnectTimeout Repmgr time before node will be fenced when insufficient child nodes are detected (in seconds) 30
postgresql.usePgRewind Use pg_rewind for standby failover (experimental) false
postgresql.audit.logHostname Add client hostnames to the log file true
postgresql.audit.logConnections Add client log-in operations to the log file false
postgresql.audit.logDisconnections Add client log-outs operations to the log file false
postgresql.audit.pgAuditLog Add operations to log using the pgAudit extension ""
postgresql.audit.pgAuditLogCatalog Log catalog using pgAudit off
postgresql.audit.clientMinMessages Message log level to share with the user error
postgresql.audit.logLinePrefix Template string for the log line prefix ""
postgresql.audit.logTimezone Timezone for the log timestamps ""
postgresql.sharedPreloadLibraries Shared preload libraries (comma-separated list) pgaudit, repmgr
postgresql.maxConnections Maximum total connections ""
postgresql.postgresConnectionLimit Maximum connections for the postgres user ""
postgresql.dbUserConnectionLimit Maximum connections for the created user ""
postgresql.tcpKeepalivesInterval TCP keepalives interval ""
postgresql.tcpKeepalivesIdle TCP keepalives idle ""
postgresql.tcpKeepalivesCount TCP keepalives count ""
postgresql.statementTimeout Statement timeout ""
postgresql.pghbaRemoveFilters Comma-separated list of patterns to remove from the pg_hba.conf file ""
postgresql.extraInitContainers Extra init containers []
postgresql.repmgrConfiguration Repmgr configuration ""
postgresql.configuration PostgreSQL configuration ""
postgresql.pgHbaConfiguration PostgreSQL client authentication configuration ""
postgresql.configurationCM Name of existing ConfigMap with configuration files ""
postgresql.extendedConf Extended PostgreSQL configuration (appended to main or default configuration). Implies volumePermissions.enabled. ""
postgresql.extendedConfCM ConfigMap with PostgreSQL extended configuration ""
postgresql.initdbScripts Dictionary of initdb scripts {}
postgresql.initdbScriptsCM ConfigMap with scripts to be run at first boot ""
postgresql.initdbScriptsSecret Secret with scripts to be run at first boot ""
postgresql.tls.enabled Enable TLS traffic support for end-client connections false
postgresql.tls.preferServerCiphers Whether to use the server's TLS cipher preferences rather than the client's true
postgresql.tls.certificatesSecret Name of an existing secret that contains the certificates ""
postgresql.tls.certFilename Certificate filename ""
postgresql.tls.certKeyFilename Certificate key filename ""
postgresql.preStopDelayAfterPgStopSeconds Minimal number of seconds preStop hook waits after postgres instance is stopped 25
postgresql.headlessWithNotReadyAddresses set postgres headless service into publishNotReadyAddresses mode false
witness.create Create PostgreSQL witness nodes false
witness.labels Labels to add to the StatefulSet. Evaluated as template {}
witness.podLabels Labels to add to the StatefulSet pods. Evaluated as template {}
witness.replicaCount Number of replicas to deploy. 1
witness.updateStrategy.type Postgresql statefulset strategy type RollingUpdate
witness.containerPorts.postgresql PostgreSQL witness port 5432
witness.hostAliases Deployment pod host aliases []
witness.hostNetwork Specify if host network should be enabled for PostgreSQL witness pod false
witness.hostIPC Specify if host IPC should be enabled for PostgreSQL witness pod false
witness.podAnnotations Additional pod annotations {}
witness.podAffinityPreset PostgreSQL witness pod affinity preset. Ignored if witness.affinity is set. Allowed values: soft or hard ""
witness.podAntiAffinityPreset PostgreSQL witness pod anti-affinity preset. Ignored if witness.affinity is set. Allowed values: soft or hard soft
witness.nodeAffinityPreset.type PostgreSQL witness node affinity preset type. Ignored if witness.affinity is set. Allowed values: soft or hard ""
witness.nodeAffinityPreset.key PostgreSQL witness node label key to match Ignored if witness.affinity is set. ""
witness.nodeAffinityPreset.values PostgreSQL witness node label values to match. Ignored if witness.affinity is set. []
witness.affinity Affinity for PostgreSQL witness pods assignment {}
witness.nodeSelector Node labels for PostgreSQL witness pods assignment {}
witness.tolerations Tolerations for PostgreSQL witness pods assignment []
witness.topologySpreadConstraints Topology Spread Constraints for pod assignment spread across your cluster among failure-domains. Evaluated as a template []
witness.priorityClassName Pod priority class ""
witness.schedulerName Use an alternate scheduler, e.g. "stork". ""
witness.terminationGracePeriodSeconds Seconds PostgreSQL witness pod needs to terminate gracefully ""
witness.podSecurityContext.enabled Enable security context for PostgreSQL witness with Repmgr true
witness.podSecurityContext.fsGroup Group ID for the PostgreSQL witness with Repmgr filesystem 1001
witness.containerSecurityContext.enabled Enable container security context true
witness.containerSecurityContext.runAsUser User ID for the PostgreSQL witness with Repmgr container 1001
witness.containerSecurityContext.runAsNonRoot Set PostgreSQL witness with Repmgr containers' Security Context runAsNonRoot true
witness.containerSecurityContext.readOnlyRootFilesystem Set PostgreSQL witness with Repmgr containers' Security Context runAsNonRoot false
witness.command Override default container command (useful when using custom images) []
witness.args Override default container args (useful when using custom images) []
witness.lifecycleHooks LifecycleHook to set additional configuration at startup, e.g. LDAP settings via REST API. Evaluated as a template {}
witness.extraEnvVars Array containing extra environment variables []
witness.extraEnvVarsCM ConfigMap with extra environment variables ""
witness.extraEnvVarsSecret Secret with extra environment variables ""
witness.extraVolumes Extra volumes to add to the deployment []
witness.extraVolumeMounts Extra volume mounts to add to the container. Normally used with extraVolumes. []
witness.initContainers Extra init containers to add to the deployment []
witness.sidecars Extra sidecar containers to add to the deployment []
witness.resources.limits The resources limits for the container {}
witness.resources.requests The requested resources for the container {}
witness.livenessProbe.enabled Enable livenessProbe true
witness.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 30
witness.livenessProbe.periodSeconds Period seconds for livenessProbe 10
witness.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
witness.livenessProbe.failureThreshold Failure threshold for livenessProbe 6
witness.livenessProbe.successThreshold Success threshold for livenessProbe 1
witness.readinessProbe.enabled Enable readinessProbe true
witness.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 5
witness.readinessProbe.periodSeconds Period seconds for readinessProbe 10
witness.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 5
witness.readinessProbe.failureThreshold Failure threshold for readinessProbe 6
witness.readinessProbe.successThreshold Success threshold for readinessProbe 1
witness.startupProbe.enabled Enable startupProbe false
witness.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 5
witness.startupProbe.periodSeconds Period seconds for startupProbe 10
witness.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
witness.startupProbe.failureThreshold Failure threshold for startupProbe 10
witness.startupProbe.successThreshold Success threshold for startupProbe 1
witness.customLivenessProbe Override default liveness probe {}
witness.customReadinessProbe Override default readiness probe {}
witness.customStartupProbe Override default startup probe {}
witness.pdb.create Specifies whether to create a Pod disruption budget for PostgreSQL witness with Repmgr false
witness.pdb.minAvailable Minimum number / percentage of pods that should remain scheduled 1
witness.pdb.maxUnavailable Maximum number / percentage of pods that may be made unavailable ""
witness.upgradeRepmgrExtension Upgrade repmgr extension in the database false
witness.pgHbaTrustAll Configures PostgreSQL HBA to trust every user false
witness.repmgrLogLevel Repmgr log level (DEBUG, INFO, NOTICE, WARNING, ERROR, ALERT, CRIT or EMERG) NOTICE
witness.repmgrConnectTimeout Repmgr backend connection timeout (in seconds) 5
witness.repmgrReconnectAttempts Repmgr backend reconnection attempts 2
witness.repmgrReconnectInterval Repmgr backend reconnection interval (in seconds) 3
witness.audit.logHostname Add client hostnames to the log file true
witness.audit.logConnections Add client log-in operations to the log file false
witness.audit.logDisconnections Add client log-outs operations to the log file false
witness.audit.pgAuditLog Add operations to log using the pgAudit extension ""
witness.audit.pgAuditLogCatalog Log catalog using pgAudit off
witness.audit.clientMinMessages Message log level to share with the user error
witness.audit.logLinePrefix Template string for the log line prefix ""
witness.audit.logTimezone Timezone for the log timestamps ""
witness.maxConnections Maximum total connections ""
witness.postgresConnectionLimit Maximum connections for the postgres user ""
witness.dbUserConnectionLimit Maximum connections for the created user ""
witness.tcpKeepalivesInterval TCP keepalives interval ""
witness.tcpKeepalivesIdle TCP keepalives idle ""
witness.tcpKeepalivesCount TCP keepalives count ""
witness.statementTimeout Statement timeout ""
witness.pghbaRemoveFilters Comma-separated list of patterns to remove from the pg_hba.conf file ""
witness.extraInitContainers Extra init containers []
witness.repmgrConfiguration Repmgr configuration ""
witness.configuration PostgreSQL configuration ""
witness.pgHbaConfiguration PostgreSQL client authentication configuration ""
witness.configurationCM Name of existing ConfigMap with configuration files ""
witness.extendedConf Extended PostgreSQL configuration (appended to main or default configuration). Implies volumePermissions.enabled. ""
witness.extendedConfCM ConfigMap with PostgreSQL extended configuration ""
witness.initdbScripts Dictionary of initdb scripts {}
witness.initdbScriptsCM ConfigMap with scripts to be run at first boot ""
witness.initdbScriptsSecret Secret with scripts to be run at first boot ""

Pgpool parameters

Name Description Value
pgpool.image.registry Pgpool image registry docker.io
pgpool.image.repository Pgpool image repository bitnami/pgpool
pgpool.image.tag Pgpool image tag 4.4.3-debian-11-r10
pgpool.image.digest Pgpool image digest in the way sha256:aa.... Please note this parameter, if set, will override the tag ""
pgpool.image.pullPolicy Pgpool image pull policy IfNotPresent
pgpool.image.pullSecrets Specify docker-registry secret names as an array []
pgpool.image.debug Specify if debug logs should be enabled false
pgpool.customUsers.usernames Comma or semicolon separated list of additional users that will be performing connections to the database using pgpool. ""
pgpool.customUsers.passwords Comma or semicolon separated list of the associated passwords for the users above. Must have the same number of elements as the usernames list. ""
pgpool.hostAliases Deployment pod host aliases []
pgpool.customUsersSecret Name of a secret containing the usernames and passwords of accounts that will be added to pgpool_passwd ""
pgpool.existingSecret Pgpool admin password using existing secret ""
pgpool.srCheckDatabase Name of the database to perform streaming replication checks postgres
pgpool.labels Labels to add to the Deployment. Evaluated as template {}
pgpool.podLabels Labels to add to the pods. Evaluated as template {}
pgpool.serviceLabels Labels to add to the service. Evaluated as template {}
pgpool.serviceAnnotations Provide any additional annotations for Pgpool service {}
pgpool.customLivenessProbe Override default liveness probe {}
pgpool.customReadinessProbe Override default readiness probe {}
pgpool.customStartupProbe Override default startup probe {}
pgpool.command Override default container command (useful when using custom images) []
pgpool.args Override default container args (useful when using custom images) []
pgpool.lifecycleHooks LifecycleHook to set additional configuration at startup, e.g. LDAP settings via REST API. Evaluated as a template {}
pgpool.extraEnvVars Array containing extra environment variables []
pgpool.extraEnvVarsCM ConfigMap with extra environment variables ""
pgpool.extraEnvVarsSecret Secret with extra environment variables ""
pgpool.extraVolumes Extra volumes to add to the deployment []
pgpool.extraVolumeMounts Extra volume mounts to add to the container. Normally used with extraVolumes []
pgpool.initContainers Extra init containers to add to the deployment []
pgpool.sidecars Extra sidecar containers to add to the deployment []
pgpool.replicaCount The number of replicas to deploy 1
pgpool.podAnnotations Additional pod annotations {}
pgpool.priorityClassName Pod priority class ""
pgpool.schedulerName Use an alternate scheduler, e.g. "stork". ""
pgpool.terminationGracePeriodSeconds Seconds pgpool pod needs to terminate gracefully ""
pgpool.topologySpreadConstraints Topology Spread Constraints for pod assignment spread across your cluster among failure-domains. Evaluated as a template []
pgpool.podAffinityPreset Pgpool pod affinity preset. Ignored if pgpool.affinity is set. Allowed values: soft or hard ""
pgpool.podAntiAffinityPreset Pgpool pod anti-affinity preset. Ignored if pgpool.affinity is set. Allowed values: soft or hard soft
pgpool.nodeAffinityPreset.type Pgpool node affinity preset type. Ignored if pgpool.affinity is set. Allowed values: soft or hard ""
pgpool.nodeAffinityPreset.key Pgpool node label key to match Ignored if pgpool.affinity is set. ""
pgpool.nodeAffinityPreset.values Pgpool node label values to match. Ignored if pgpool.affinity is set. []
pgpool.affinity Affinity for Pgpool pods assignment {}
pgpool.nodeSelector Node labels for Pgpool pods assignment {}
pgpool.tolerations Tolerations for Pgpool pods assignment []
pgpool.podSecurityContext.enabled Enable security context for Pgpool true
pgpool.podSecurityContext.fsGroup Group ID for the Pgpool filesystem 1001
pgpool.containerSecurityContext.enabled Enable container security context true
pgpool.containerSecurityContext.runAsUser User ID for the Pgpool container 1001
pgpool.containerSecurityContext.runAsNonRoot Set Pgpool containers' Security Context runAsNonRoot true
pgpool.containerSecurityContext.readOnlyRootFilesystem Set Pgpool containers' Security Context runAsNonRoot false
pgpool.resources.limits The resources limits for the container {}
pgpool.resources.requests The requested resources for the container {}
pgpool.livenessProbe.enabled Enable livenessProbe true
pgpool.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 30
pgpool.livenessProbe.periodSeconds Period seconds for livenessProbe 10
pgpool.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
pgpool.livenessProbe.failureThreshold Failure threshold for livenessProbe 5
pgpool.livenessProbe.successThreshold Success threshold for livenessProbe 1
pgpool.readinessProbe.enabled Enable readinessProbe true
pgpool.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 5
pgpool.readinessProbe.periodSeconds Period seconds for readinessProbe 5
pgpool.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 5
pgpool.readinessProbe.failureThreshold Failure threshold for readinessProbe 5
pgpool.readinessProbe.successThreshold Success threshold for readinessProbe 1
pgpool.startupProbe.enabled Enable startupProbe false
pgpool.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 5
pgpool.startupProbe.periodSeconds Period seconds for startupProbe 10
pgpool.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
pgpool.startupProbe.failureThreshold Failure threshold for startupProbe 10
pgpool.startupProbe.successThreshold Success threshold for startupProbe 1
pgpool.pdb.create Specifies whether a Pod disruption budget should be created for Pgpool pods false
pgpool.pdb.minAvailable Minimum number / percentage of pods that should remain scheduled 1
pgpool.pdb.maxUnavailable Maximum number / percentage of pods that may be made unavailable ""
pgpool.updateStrategy Strategy used to replace old Pods by new ones {}
pgpool.containerPorts.postgresql Pgpool port 5432
pgpool.minReadySeconds How many seconds a pod needs to be ready before killing the next, during update ""
pgpool.adminUsername Pgpool Admin username admin
pgpool.adminPassword Pgpool Admin password ""
pgpool.usePasswordFile Set to true to mount pgpool secret as a file instead of passing environment variable ""
pgpool.authenticationMethod Pgpool authentication method. Use 'md5' for PSQL < 14. scram-sha-256
pgpool.logConnections Log all client connections (PGPOOL_ENABLE_LOG_CONNECTIONS) false
pgpool.logHostname Log the client hostname instead of IP address (PGPOOL_ENABLE_LOG_HOSTNAME) true
pgpool.logPerNodeStatement Log every SQL statement for each DB node separately (PGPOOL_ENABLE_LOG_PER_NODE_STATEMENT) false
pgpool.logLinePrefix Format of the log entry lines (PGPOOL_LOG_LINE_PREFIX) ""
pgpool.clientMinMessages Log level for clients error
pgpool.numInitChildren The number of preforked Pgpool-II server processes. It is also the concurrent ""
pgpool.reservedConnections Number of reserved connections. When zero, excess connection block. When non-zero, excess connections are refused with an error message. 1
pgpool.maxPool The maximum number of cached connections in each child process (PGPOOL_MAX_POOL) ""
pgpool.childMaxConnections The maximum number of client connections in each child process (PGPOOL_CHILD_MAX_CONNECTIONS) ""
pgpool.childLifeTime The time in seconds to terminate a Pgpool-II child process if it remains idle (PGPOOL_CHILD_LIFE_TIME) ""
pgpool.clientIdleLimit The time in seconds to disconnect a client if it remains idle since the last query (PGPOOL_CLIENT_IDLE_LIMIT) ""
pgpool.connectionLifeTime The time in seconds to terminate the cached connections to the PostgreSQL backend (PGPOOL_CONNECTION_LIFE_TIME) ""
pgpool.useLoadBalancing Use Pgpool Load-Balancing true
pgpool.loadBalancingOnWrite LoadBalancer on write actions behavior transaction
pgpool.configuration Pgpool configuration ""
pgpool.configurationCM ConfigMap with Pgpool configuration ""
pgpool.initdbScripts Dictionary of initdb scripts {}
pgpool.initdbScriptsCM ConfigMap with scripts to be run every time Pgpool container is initialized ""
pgpool.initdbScriptsSecret Secret with scripts to be run every time Pgpool container is initialized ""
pgpool.tls.enabled Enable TLS traffic support for end-client connections false
pgpool.tls.autoGenerated Create self-signed TLS certificates. Currently only supports PEM certificates false
pgpool.tls.preferServerCiphers Whether to use the server's TLS cipher preferences rather than the client's true
pgpool.tls.certificatesSecret Name of an existing secret that contains the certificates ""
pgpool.tls.certFilename Certificate filename ""
pgpool.tls.certKeyFilename Certificate key filename ""
pgpool.tls.certCAFilename CA Certificate filename ""

LDAP parameters

Name Description Value
ldap.enabled Enable LDAP support false
ldap.existingSecret Name of existing secret to use for LDAP passwords ""
ldap.uri LDAP URL beginning in the form ldap[s]://<hostname>:<port> ""
ldap.basedn LDAP base DN ""
ldap.binddn LDAP bind DN ""
ldap.bindpw LDAP bind password ""
ldap.bslookup LDAP base lookup ""
ldap.scope LDAP search scope ""
ldap.tlsReqcert LDAP TLS check on server certificates ""
ldap.nssInitgroupsIgnoreusers LDAP ignored users root,nslcd

Other Parameters

Name Description Value
rbac.create Create Role and RoleBinding (required for PSP to work) false
rbac.rules Custom RBAC rules to set []
serviceAccount.create Enable creation of ServiceAccount for Airflow pods false
serviceAccount.name The name of the ServiceAccount to use. ""
serviceAccount.annotations Additional custom annotations for the ServiceAccount {}
serviceAccount.automountServiceAccountToken Allows auto mount of ServiceAccountToken on the serviceAccount created true
psp.create Whether to create a PodSecurityPolicy. WARNING: PodSecurityPolicy is deprecated in Kubernetes v1.21 or later, unavailable in v1.25 or later false

Metrics parameters

Name Description Value
metrics.enabled Enable PostgreSQL Prometheus exporter false
metrics.image.registry PostgreSQL Prometheus exporter image registry docker.io
metrics.image.repository PostgreSQL Prometheus exporter image repository bitnami/postgres-exporter
metrics.image.tag PostgreSQL Prometheus exporter image tag 0.13.1-debian-11-r0
metrics.image.digest PostgreSQL Prometheus exporter image digest in the way sha256:aa.... Please note this parameter, if set, will override the tag ""
metrics.image.pullPolicy PostgreSQL Prometheus exporter image pull policy IfNotPresent
metrics.image.pullSecrets Specify docker-registry secret names as an array []
metrics.image.debug Specify if debug logs should be enabled false
metrics.podSecurityContext.enabled Enable security context for PostgreSQL Prometheus exporter true
metrics.podSecurityContext.runAsUser User ID for the PostgreSQL Prometheus exporter container 1001
metrics.resources.limits The resources limits for the container {}
metrics.resources.requests The requested resources for the container {}
metrics.containerPorts.http Prometheus metrics exporter port 9187
metrics.livenessProbe.enabled Enable livenessProbe true
metrics.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 30
metrics.livenessProbe.periodSeconds Period seconds for livenessProbe 10
metrics.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
metrics.livenessProbe.failureThreshold Failure threshold for livenessProbe 6
metrics.livenessProbe.successThreshold Success threshold for livenessProbe 1
metrics.readinessProbe.enabled Enable readinessProbe true
metrics.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 5
metrics.readinessProbe.periodSeconds Period seconds for readinessProbe 10
metrics.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 5
metrics.readinessProbe.failureThreshold Failure threshold for readinessProbe 6
metrics.readinessProbe.successThreshold Success threshold for readinessProbe 1
metrics.startupProbe.enabled Enable startupProbe false
metrics.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 5
metrics.startupProbe.periodSeconds Period seconds for startupProbe 10
metrics.startupProbe.timeoutSeconds Timeout seconds for startupProbe 5
metrics.startupProbe.failureThreshold Failure threshold for startupProbe 10
metrics.startupProbe.successThreshold Success threshold for startupProbe 1
metrics.customLivenessProbe Override default liveness probe {}
metrics.customReadinessProbe Override default readiness probe {}
metrics.customStartupProbe Override default startup probe {}
metrics.service.type PostgreSQL Prometheus exporter metrics service type ClusterIP
metrics.service.ports.metrics PostgreSQL Prometheus exporter metrics service port 9187
metrics.service.nodePorts.metrics PostgreSQL Prometheus exporter Node Port ""
metrics.service.clusterIP PostgreSQL Prometheus exporter metrics service Cluster IP ""
metrics.service.loadBalancerIP PostgreSQL Prometheus exporter service Load Balancer IP ""
metrics.service.loadBalancerSourceRanges PostgreSQL Prometheus exporter service Load Balancer sources []
metrics.service.externalTrafficPolicy PostgreSQL Prometheus exporter service external traffic policy Cluster
metrics.annotations Annotations for PostgreSQL Prometheus exporter service {}
metrics.customMetrics Additional custom metrics {}
metrics.extraEnvVars Array containing extra environment variables []
metrics.extraEnvVarsCM ConfigMap with extra environment variables ""
metrics.extraEnvVarsSecret Secret with extra environment variables ""
metrics.serviceMonitor.enabled if true, creates a Prometheus Operator ServiceMonitor (also requires metrics.enabled to be true) false
metrics.serviceMonitor.namespace Optional namespace which Prometheus is running in ""
metrics.serviceMonitor.interval How frequently to scrape metrics (use by default, falling back to Prometheus' default) ""
metrics.serviceMonitor.scrapeTimeout Service monitor scrape timeout ""
metrics.serviceMonitor.annotations Additional annotations for the ServiceMonitor {}
metrics.serviceMonitor.labels Additional labels that can be used so ServiceMonitor will be discovered by Prometheus {}
metrics.serviceMonitor.selector Defaults to what's used if you follow CoreOS Prometheus Install Instructions (https://github.com/bitnami/charts/tree/main/bitnami/kube-prometheus#tldr) {}
metrics.serviceMonitor.relabelings ServiceMonitor relabelings. Value is evaluated as a template []
metrics.serviceMonitor.metricRelabelings ServiceMonitor metricRelabelings. Value is evaluated as a template []
metrics.serviceMonitor.honorLabels Specify honorLabels parameter to add the scrape endpoint false
metrics.serviceMonitor.jobLabel The name of the label on the target service to use as the job name in prometheus. ""

Volume permissions parameters

Name Description Value
volumePermissions.enabled Enable init container to adapt volume permissions false
volumePermissions.image.registry Init container volume-permissions image registry docker.io
volumePermissions.image.repository Init container volume-permissions image repository bitnami/bitnami-shell
volumePermissions.image.tag Init container volume-permissions image tag 11-debian-11-r130
volumePermissions.image.digest Init container volume-permissions image digest in the way sha256:aa.... Please note this parameter, if set, will override the tag ""
volumePermissions.image.pullPolicy Init container volume-permissions image pull policy IfNotPresent
volumePermissions.image.pullSecrets Specify docker-registry secret names as an array []
volumePermissions.podSecurityContext.runAsUser Init container volume-permissions User ID 0
volumePermissions.resources.limits The resources limits for the container {}
volumePermissions.resources.requests The requested resources for the container {}

Persistence parameters

Name Description Value
persistence.enabled Enable data persistence true
persistence.existingClaim A manually managed Persistent Volume and Claim ""
persistence.storageClass Persistent Volume Storage Class ""
persistence.mountPath The path the volume will be mounted at, useful when using different PostgreSQL images. /bitnami/postgresql
persistence.accessModes List of access modes of data volume ["ReadWriteOnce"]
persistence.size Persistent Volume Claim size 8Gi
persistence.annotations Persistent Volume Claim annotations {}
persistence.labels Persistent Volume Claim labels {}
persistence.selector Selector to match an existing Persistent Volume (this value is evaluated as a template) {}

Traffic Exposure parameters

Name Description Value
service.type Kubernetes service type (ClusterIP, NodePort or LoadBalancer) ClusterIP
service.ports.postgresql PostgreSQL port 5432
service.portName PostgreSQL service port name postgresql
service.nodePorts.postgresql Kubernetes service nodePort ""
service.loadBalancerIP Load balancer IP if service type is LoadBalancer ""
service.loadBalancerSourceRanges Addresses that are allowed when service is LoadBalancer []
service.clusterIP Set the Cluster IP to use ""
service.externalTrafficPolicy Enable client source IP preservation Cluster
service.extraPorts Extra ports to expose (normally used with the sidecar value) []
service.sessionAffinity Control where client requests go, to the same pod or round-robin None
service.sessionAffinityConfig Additional settings for the sessionAffinity {}
service.annotations Provide any additional annotations both for PostgreSQL and Pgpool services {}
service.serviceLabels Labels for PostgreSQL service {}
service.headless.annotations Annotations for the headless service. {}
networkPolicy.enabled Enable NetworkPolicy false
networkPolicy.allowExternal Don't require client label for connections true
networkPolicy.egressRules.denyConnectionsToExternal Enable egress rule that denies outgoing traffic outside the cluster, except for DNS (port 53) false
networkPolicy.egressRules.customRules Custom network policy rule {}

Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

helm install my-release \
    --set postgresql.password=password \
    bitnami/postgresql-ha

The above command sets the password for user postgres to password.

NOTE: Once this chart is deployed, it is not possible to change the application's access credentials, such as usernames or passwords, using Helm. To change these application credentials after deployment, delete any persistent volumes (PVs) used by the chart and re-deploy it, or use the application's built-in administrative tools if available.

Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,

helm install my-release \
    -f values.yaml \
    bitnami/postgresql-ha

Configuration and installation details

Rolling VS Immutable tags

It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.

Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.

Use a different PostgreSQL version

To modify the application version used in this chart, specify a different version of the image using the image.tag parameter and/or a different repository using the image.repository parameter. Refer to the chart documentation for more information on these parameters and how to use them with images from a private registry.

Use a volume for /dev/shm

When working with huge databeses, /dev/shm can run out of space. A way to fix this is to use the postgresql.extraVolumes and postgresql.extraVolumeMounts values. In the example below, we set an emptyDir volume with 512Mb:

postgresql:
  extraVolumes:
    - name: dshm
      emptyDir:
        medium: Memory
        sizeLimit: 512Mi
  extraVolumeMounts:
    - name: dshm
      mountPath: /dev/shm

Configure the way how to expose PostgreSQL

  • ClusterIP: Exposes the service on a cluster-internal IP. Choosing this value makes the service only reachable from within the cluster. Set service.type=ClusterIP to choose this service type.
  • NodePort: Exposes the service on each Node's IP at a static port (the NodePort). You will be able to contact the NodePort service, from outside the cluster, by requesting NodeIP:NodePort. Set service.type=NodePort to choose this service type.
  • LoadBalancer: Exposes the service externally using a cloud provider's load balancer. Set service.type=LoadBalancer to choose this service type.

Adjust permissions of persistent volume mountpoint

As the images run as non-root by default, it is necessary to adjust the ownership of the persistent volumes so that the containers can write data into it.

By default, the chart is configured to use Kubernetes Security Context to automatically change the ownership of the volume. However, this feature does not work in all Kubernetes distributions. As an alternative, this chart supports using an initContainer to change the ownership of the volume before mounting it in the final destination.

You can enable this initContainer by setting volumePermissions.enabled to true.

Securing traffic using TLS

Learn how to configure TLS authentication

LDAP

LDAP support can be enabled in the chart by specifying the ldap. parameters while creating a release. The following parameters should be configured to properly enable the LDAP support in the chart.

  • ldap.enabled: Enable LDAP support. Defaults to false.
  • ldap.uri: LDAP URL beginning in the form ldap[s]://<hostname>:<port>. No defaults.
  • ldap.base: LDAP base DN. No defaults.
  • ldap.binddn: LDAP bind DN. No defaults.
  • ldap.bindpw: LDAP bind password. No defaults.
  • ldap.bslookup: LDAP base lookup. No defaults.
  • ldap.nss_initgroups_ignoreusers: LDAP ignored users. root,nslcd.
  • ldap.scope: LDAP search scope. No defaults.
  • ldap.tls_reqcert: LDAP TLS check on server certificates. No defaults.

For example:

ldap.enabled="true"
ldap.uri="ldap://my_ldap_server"
ldap.base="dc=example\,dc=org"
ldap.binddn="cn=admin\,dc=example\,dc=org"
ldap.bindpw="admin"
ldap.bslookup="ou=group-ok\,dc=example\,dc=org"
ldap.nss_initgroups_ignoreusers="root\,nslcd"
ldap.scope="sub"
ldap.tls_reqcert="demand"

Next, login to the PostgreSQL server using the psql client and add the PAM authenticated LDAP users.

Note: Parameters including commas must be escaped as shown in the above example.

repmgr.conf / postgresql.conf / pg_hba.conf / pgpool.conf files as configMap

This helm chart also supports to customize the whole configuration file.

You can specify the Pgpool, PostgreSQL and Repmgr configuration using the pgpool.configuration, postgresql.configuration, postgresql.pgHbaConfiguration, and postgresql.repmgrConfiguration parameters. The corresponding files will be mounted as ConfigMap to the containers and it will be used for configuring Pgpool, Repmgr and the PostgreSQL server.

In addition to this option, you can also set an external ConfigMap(s) with all the configuration files. This is done by setting the postgresql.configurationCM and pgpool.configurationCM parameters. Note that this will override the previous options.

Allow settings to be loaded from files other than the default postgresql.conf

If you don't want to provide the whole PostgreSQL configuration file and only specify certain parameters, you can specify the extended configuration using the postgresql.extendedConf parameter. A file will be mounted as configMap to the containers adding/overwriting the default configuration using the include_dir directive that allows settings to be loaded from files other than the default postgresql.conf.

In addition to this option, you can also set an external ConfigMap with all the extra configuration files. This is done by setting the postgresql.extendedConfCM parameter. Note that this will override the previous option.

Initialize a fresh instance

The Bitnami PostgreSQL with Repmgr image allows you to use your custom scripts to initialize a fresh instance. You can specify custom scripts using the initdbScripts parameter as dict so they can be consumed as a ConfigMap.

In addition to this option, you can also set an external ConfigMap with all the initialization scripts. This is done by setting the initdbScriptsCM parameter. Note that this will override the two previous options. If your initialization scripts contain sensitive information such as credentials or passwords, you can use the initdbScriptsSecret parameter.

The above parameters (initdbScripts, initdbScriptsCM, and initdbScriptsSecret) are supported in both StatefulSet by prepending postgresql or pgpool to the parameter, depending on the use case (see above parameters table).

The allowed extensions are .sh, .sql and .sql.gz in the postgresql container while only .sh in the case of the pgpool one.

+info: https://github.com/bitnami/containers/tree/main/bitnami/postgresql#initializing-a-new-instance and https://github.com/bitnami/containers/tree/main/bitnami/pgpool#initializing-with-custom-scripts

Use of global variables

In more complex scenarios, we may have the following tree of dependencies

                     +--------------+
                     |              |
        +------------+   Chart 1    +-----------+
        |            |              |           |
        |            --------+------+           |
        |                    |                  |
        |                    |                  |
        |                    |                  |
        |                    |                  |
        v                    v                  v
+-------+------+    +--------+------+  +--------+------+
|              |    |               |  |               |
| PostgreSQL HA |  | Sub-chart 1 |  | Sub-chart 2 |
|---------------|--|-------------|--|-------------|
+--------------+    +---------------+  +---------------+

The three charts below depend on the parent chart Chart 1. However, subcharts 1 and 2 may need to connect to PostgreSQL HA as well. In order to do so, subcharts 1 and 2 need to know the PostgreSQL HA credentials, so one option for deploying could be deploy Chart 1 with the following parameters:

postgresql.postgresqlPassword=testtest
subchart1.postgresql.postgresqlPassword=testtest
subchart2.postgresql.postgresqlPassword=testtest
postgresql.postgresqlDatabase=db1
subchart1.postgresql.postgresqlDatabase=db1
subchart2.postgresql.postgresqlDatabase=db1

If the number of dependent sub-charts increases, installing the chart with parameters can become increasingly difficult. An alternative would be to set the credentials using global variables as follows:

global.postgresql.postgresqlPassword=testtest
global.postgresql.postgresqlDatabase=db1

This way, the credentials will be available in all of the subcharts.

Persistence

The data is persisted by default using PVC templates in the PostgreSQL statefulset. You can disable the persistence setting the persistence.enabled parameter to false. A default StorageClass is needed in the Kubernetes cluster to dynamically provision the volumes. Specify another StorageClass in the persistence.storageClass or set persistence.existingClaim if you have already existing persistent volumes to use.

Setting Pod's affinity

This chart allows you to set your custom affinity using the XXX.affinity paremeter(s). Find more information about Pod's affinity in the kubernetes documentation.

As an alternative, you can use of the preset configurations for pod affinity, pod anti-affinity, and node affinity available at the bitnami/common chart. To do so, set the XXX.podAffinityPreset, XXX.podAntiAffinityPreset, or XXX.nodeAffinityPreset parameters.

Troubleshooting

Find more information about how to deal with common errors related to Bitnami's Helm charts in this troubleshooting guide.

Upgrading

It's necessary to specify the existing passwords while performing a upgrade to ensure the secrets are not updated with invalid randomly generated passwords. Remember to specify the existing values of the postgresql.password and postgresql.repmgrPassword parameters when upgrading the chart:

helm upgrade my-release bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRES_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD]

Note: you need to substitute the placeholders [POSTGRES_PASSWORD], and [REPMGR_PASSWORD] with the values obtained from instructions in the installation notes. Note: As general rule, it is always wise to do a backup before the upgrading procedures.

To 10.0.0

This major version changes the default PostgreSQL image from 14.x to 15.x. Follow the official instructions to upgrade to 15.x.

To 9.0.0

This chart major version updates the PostgreSQL image's version to the latest major, v14, as well as standarizes the templates and values. These changes can be sumarised in the following:

  • Image parameters that used imageNameImage are now under imageName.image
  • containerPort parameters are now found by containerPorts.xxxx
  • service.port parameters are now found by service.ports.xxxx
  • imageName.securityContext parameters are now under imageName.podSecurityContext
  • serviceAccount.enabled parameter has been renamed to serviceAccount.create
  • postgresql.updateStrategyType parameter has been renamed to postgresql.updateStrategy.type
  • Added the pgpool.authenticationMethod parameter. Use "md5" for older PSQL versions and the default value of "scram-sha-256" when using PSQL > v14.

Note: Keep in mind you will find an error such as the one below when upgrading since the new chart major version also bumps the application version. To workaround this issue you need to upgrade the database, please refer to the official PostgreSQL documentation for more information about this.

postgresql-repmgr 14:41:21.00 INFO  ==> Setting up streaming replication slave...
postgresql-repmgr 14:41:21.04 INFO  ==> Starting PostgreSQL in background...
waiting for server to start....2022-04-12 14:41:21.069 GMT [220] FATAL:  database files are incompatible with server
2022-04-12 14:41:21.069 GMT [220] DETAIL:  The data directory was initialized by PostgreSQL version 11, which is not compatible with this version 14.2.
pg_ctl: could not start server

To 8.0.0

A new major version of repmgr (5.3) was included. To upgrade to this major version, it's necessary to upgrade the repmgr extension installed on the database. To do so, follow the steps below:

  • Reduce your PostgreSQL setup to one replica (primary node) and upgrade to 8.0.0, enabling the repmgr extension upgrade:
helm upgrade my-release --version 8.0.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=1 \
    --set postgresql.upgradeRepmgrExtension=true
  • Scale your PostgreSQL setup to the original number of replicas:
helm upgrade my-release --version 8.0.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=[NUMBER_OF_REPLICAS]

Note: you need to substitute the placeholders [POSTGRESQL_PASSWORD], and [REPMGR_PASSWORD] with the values obtained from instructions in the installation notes.

To 6.4.0

Support for adding custom configuration files or initialization scripts by placing them under the "files" directory in the working directory was removed. This functionality was very confusing for users since they do not usually clone the repo nor they fetch the charts to their working directories. As an alternative to this feature, users can still use the equivalent parameters available in the values.yaml to load their custom configuration & scripts.

To 6.0.0

On November 13, 2020, Helm v2 support was formally finished, this major version is the result of the required changes applied to the Helm Chart to be able to incorporate the different features added in Helm v3 and to be consistent with the Helm project itself regarding the Helm v2 EOL.

What changes were introduced in this major version?

  • Previous versions of this Helm Chart use apiVersion: v1 (installable by both Helm 2 and 3), this Helm Chart was updated to apiVersion: v2 (installable by Helm 3 only). Here you can find more information about the apiVersion field.
  • Move dependency information from the requirements.yaml to the Chart.yaml
  • After running helm dependency update, a Chart.lock file is generated containing the same structure used in the previous requirements.lock
  • The different fields present in the Chart.yaml file has been ordered alphabetically in a homogeneous way for all the Bitnami Helm Charts

Considerations when upgrading to this version

  • If you want to upgrade to this version from a previous one installed with Helm v3, you shouldn't face any issues
  • If you want to upgrade to this version using Helm v2, this scenario is not supported as this version doesn't support Helm v2 anymore
  • If you installed the previous version with Helm v2 and wants to upgrade to this version with Helm v3, please refer to the official Helm documentation about migrating from Helm v2 to v3

To 5.2.0

A new version of repmgr (5.2.0) was included. To upgrade to this version, it's necessary to upgrade the repmgr extension installed on the database. To do so, follow the steps below:

  • Reduce your PostgreSQL setup to one replica (primary node) and upgrade to 5.2.0, enabling the repmgr extension upgrade:
helm upgrade my-release --version 5.2.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=1 \
    --set postgresql.upgradeRepmgrExtension=true
  • Scale your PostgreSQL setup to the original number of replicas:
helm upgrade my-release --version 5.2.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=[NUMBER_OF_REPLICAS]

Note: you need to substitute the placeholders [POSTGRESQL_PASSWORD], and [REPMGR_PASSWORD] with the values obtained from instructions in the installation notes (helm get notes RELEASE_NAME).

To 5.0.0

This release uses parallel deployment for the postgresql statefullset. This should fix the issues related to not being able to restart the cluster under some contions where the master node is not longer node -0. This version is next major version to v3.x.y

  • To upgrade to this version you will need to delete the deployment, keep the PVCs and launch a new deployment keeping the deployment name.
$ # e.g. Previous deployment v3.9.1
helm install my-release \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    bitnami/postgresql-ha --version 3.9.1

$ # Update repository information
helm repo update

$ # upgrade to v5.0.0
helm delete my-release
helm install my-release \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    bitnami/postgresql-ha --version 5.0.0

To 4.0.x

Due to an error handling the version numbers these versions are actually part of the 3.x versions and not a new major version.

To 3.0.0

A new major version of repmgr (5.1.0) was included. To upgrade to this major version, it's necessary to upgrade the repmgr extension installed on the database. To do so, follow the steps below:

  • Reduce your PostgreSQL setup to one replica (primary node) and upgrade to 3.0.0, enabling the repmgr extension upgrade:
helm upgrade my-release --version 3.0.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=1 \
    --set postgresql.upgradeRepmgrExtension=true
  • Scale your PostgreSQL setup to the original number of replicas:
helm upgrade my-release --version 3.0.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=[NUMBER_OF_REPLICAS]

Note: you need to substitute the placeholders [POSTGRESQL_PASSWORD], and [REPMGR_PASSWORD] with the values obtained from instructions in the installation notes.

To 2.0.0

The Bitnami Pgpool image was migrated to a "non-root" user approach. Previously the container ran as the root user and the Pgpool daemon was started as the pgpool user. From now on, both the container and the Pgpool daemon run as user 1001. You can revert this behavior by setting the parameters pgpool.containerSecurityContext.runAsUser, and pgpool.securityContext.fsGroup to 0.

Consequences:

  • No backwards compatibility issues are expected since all the data is at PostgreSQL pods, and Pgpool uses a deployment without persistence. Therefore, upgrades should work smoothly from 1.x.x versions.
  • Environment variables related to LDAP configuration were renamed removing the PGPOOL_ prefix. For instance, to indicate the LDAP URI to use, you must set LDAP_URI instead of PGPOOL_LDAP_URI

To 1.0.0

A new major version of repmgr (5.0.0) was included. To upgrade to this major version, it's necessary to upgrade the repmgr extension installed on the database. To do so, follow the steps below:

  • Reduce your PostgreSQL setup to one replica (primary node) and upgrade to 1.0.0, enabling the repmgr extension upgrade:
helm upgrade my-release --version 1.0.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=1 \
    --set postgresql.upgradeRepmgrExtension=true
  • Scale your PostgreSQL setup to the original number of replicas:
helm upgrade my-release --version 1.0.0 bitnami/postgresql-ha \
    --set postgresql.password=[POSTGRESQL_PASSWORD] \
    --set postgresql.repmgrPassword=[REPMGR_PASSWORD] \
    --set postgresql.replicaCount=[NUMBER_OF_REPLICAS]

Note: you need to substitute the placeholders [POSTGRESQL_PASSWORD], and [REPMGR_PASSWORD] with the values obtained from instructions in the installation notes.

To 0.4.0

In this version, the chart will use PostgreSQL-Repmgr container images with the Postgis extension included. The version used in Postgresql version 10, 11 and 12 is Postgis 2.5, and in Postgresql 9.6 is Postgis 2.3. Postgis has been compiled with the following dependencies:

  • protobuf
  • protobuf-c
  • json-c
  • geos
  • proj
  • gdal

Bitnami Kubernetes Documentation

Bitnami Kubernetes documentation is available at https://docs.bitnami.com/. You can find there the following resources:

License

Copyright © 2023 VMware, Inc.

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.