++ echo 'Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/logs/scaling.log' Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/logs/scaling.log ++ '[' -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP= ++ oc get projects error: the server doesn't have a resource type "projects" +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep '\-eks\-' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ '[' ']' +++ kubectl version -o json +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ /usr/bin/sed -r 's/[^0-9.]+//g' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ KUBE_VERSION=1.26 +++ helm version -c +++ /usr/bin/sed -re 's/.*SemVer:"([^"]+)".*/\1/; s/.*\bVersion:"([^"]+)".*/\1/' ++ HELM_VERSION=v3.15.1 ++ '[' v3 == v2 ']' + create_namespace scaling-8587 + local namespace=scaling-8587 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + cleanup_rbac + awk '{print$1}' + xargs kubectl delete ns + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp + local LAST_OUT=/tmp/tmp.29xfCfmOzn ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.wXgEAftIgu ++ mktemp + local LAST_ERR=/tmp/tmp.fip26DAnXQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + local LAST_ERR=/tmp/tmp.pePhJaQVsR + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get ns + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.29xfCfmOzn + cat /tmp/tmp.fip26DAnXQ + rm /tmp/tmp.29xfCfmOzn /tmp/tmp.fip26DAnXQ + return 0 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level namespace "init-deploy-18965" deleted + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.wXgEAftIgu + cat /tmp/tmp.pePhJaQVsR Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.wXgEAftIgu /tmp/tmp.pePhJaQVsR + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp + local LAST_OUT=/tmp/tmp.hxeqZ8E3eX ++ mktemp + local LAST_ERR=/tmp/tmp.gGKCo9LaZr + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.hxeqZ8E3eX + cat /tmp/tmp.gGKCo9LaZr Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.hxeqZ8E3eX /tmp/tmp.gGKCo9LaZr + return 1 + true + kubectl delete MutatingWebhookConfiguration/chaos-mesh-mutation ValidatingWebhookConfiguration/chaos-mesh-validation ValidatingWebhookConfiguration/validate-auth Error from server (NotFound): mutatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-mutation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-validation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "validate-auth" not found + true + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin delete namespace scaling-8587 ++ mktemp + local LAST_OUT=/tmp/tmp.uXLGMcBm2J ++ mktemp + local LAST_ERR=/tmp/tmp.WCDl4gfhGK + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete namespace scaling-8587 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete namespace scaling-8587 + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete namespace scaling-8587 + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.uXLGMcBm2J + cat /tmp/tmp.WCDl4gfhGK Error from server (NotFound): namespaces "scaling-8587" not found + rm /tmp/tmp.uXLGMcBm2J /tmp/tmp.WCDl4gfhGK + return 1 + : + wait_for_delete namespace/scaling-8587 + local res=namespace/scaling-8587 + set +o xtrace namespace/scaling-8587 - Error from server (NotFound): namespaces "scaling-8587" not found + kubectl_bin create namespace scaling-8587 ++ mktemp + local LAST_OUT=/tmp/tmp.GW0NBAawOC ++ mktemp + local LAST_ERR=/tmp/tmp.rXak1dujGf + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl create namespace scaling-8587 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.GW0NBAawOC namespace/scaling-8587 created + cat /tmp/tmp.rXak1dujGf + rm /tmp/tmp.GW0NBAawOC /tmp/tmp.rXak1dujGf + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.dK3w1p7ws5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.i8UOASbmjm ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.dK3w1p7ws5 ++ cat /tmp/tmp.i8UOASbmjm ++ rm /tmp/tmp.dK3w1p7ws5 /tmp/tmp.i8UOASbmjm ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-sandbox --namespace=scaling-8587 ++ mktemp + local LAST_OUT=/tmp/tmp.CHWF6BRHve ++ mktemp + local LAST_ERR=/tmp/tmp.aQgWJVhS7c + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-sandbox --namespace=scaling-8587 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.CHWF6BRHve Context "gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-sandbox" modified. + cat /tmp/tmp.aQgWJVhS7c + rm /tmp/tmp.CHWF6BRHve /tmp/tmp.aQgWJVhS7c + return 0 + deploy_operator + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml + local namespace_mode=disabled + local action=install + local namespace=scaling-8587 + local dont_send_telemetry=false + local version_service_uri_override= + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + prepare_operator_yaml '' /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml disabled install scaling-8587 false + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml + local namespace_mode=disabled + local operator_action=install + local namespace=scaling-8587 + local dont_send_telemetry=false + yq r -d2 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml 'data[values.yaml]' + /usr/bin/sed -e 's#^namespace: .*#namespace: "scaling-8587"#g' + /usr/bin/sed -e 's#pgo_operator_namespace: .*#pgo_operator_namespace: "scaling-8587"#g' + /usr/bin/sed -e 's#namespace_mode: .*#namespace_mode: "disabled"#g' + cat /dev/fd/63 - ++ echo 'crunchy_debug: "true"' + /usr/bin/sed -e 's#^disable_telemetry: .*#disable_telemetry: "false"#g' ++ echo perconalab/percona-postgresql-operator:PR-776-0087ef129 ++ cut -d: -f2 ++ echo perconalab/percona-postgresql-operator:PR-776-0087ef129 ++ cut -d: -f1 + /usr/bin/sed -e 's#pgo_image_tag: .*#pgo_image_tag: "PR-776-0087ef129"#g' + /usr/bin/sed -e 's#pgo_image_prefix: .*#pgo_image_prefix: "perconalab/percona-postgresql-operator"#g' + [[ -n '' ]] + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + yq w '-d*' /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml metadata.namespace scaling-8587 + yq w -d3 - 'subjects[0].namespace' scaling-8587 + yq w -d4 - 'spec.template.spec.containers[0].image' perconalab/percona-postgresql-operator:PR-776-0087ef129-pgo-deployer + yq w -d4 - 'spec.template.spec.containers[0].env[0].value' install ++ cat /tmp/tmp.WATPD0tkZe/operator.ini + yq w -d2 - -d2 'data[values.yaml]' 'crunchy_debug: "true" archive_mode: "true" archive_timeout: "60" ccp_image_pull_secret: "" ccp_image_pull_secret_manifest: "" create_rbac: "true" delete_operator_namespace: "false" delete_watched_namespaces: "false" disable_telemetry: "false" namespace: "scaling-8587" namespace_mode: "disabled" pgo_image_prefix: "perconalab/percona-postgresql-operator" pgo_image_pull_policy: "Always" pgo_image_pull_secret: "" pgo_image_pull_secret_manifest: "" pgo_image_tag: "PR-776-0087ef129" pgo_operator_namespace: "scaling-8587"' + [[ -n '' ]] + kubectl_bin apply -f /tmp/tmp.WATPD0tkZe/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.8FYTga9LDn ++ mktemp + local LAST_ERR=/tmp/tmp.9m8m5fdMJi + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.WATPD0tkZe/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.8FYTga9LDn serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb configured job.batch/pgo-deploy created + cat /tmp/tmp.9m8m5fdMJi + rm /tmp/tmp.8FYTga9LDn /tmp/tmp.9m8m5fdMJi + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=scaling-8587 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 15 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 16 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 17 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 18 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 19 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 20 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 21 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 22 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 23 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 24 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 25 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 26 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 27 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 28 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 29 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 30 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 31 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.WATPD0tkZe/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.JGtNtTzN1L ++ mktemp + local LAST_ERR=/tmp/tmp.OHBFykkGHD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.WATPD0tkZe/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.JGtNtTzN1L serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.OHBFykkGHD + rm /tmp/tmp.JGtNtTzN1L /tmp/tmp.OHBFykkGHD + return 0 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uqMcUJcF9o +++ mktemp ++ local LAST_ERR=/tmp/tmp.69mQ1gu5ST ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.uqMcUJcF9o ++ cat /tmp/tmp.69mQ1gu5ST ++ rm /tmp/tmp.uqMcUJcF9o /tmp/tmp.69mQ1gu5ST ++ return 0 + wait_pod postgres-operator-778d48f559-2czmn + local pod=postgres-operator-778d48f559-2czmn + local ns= + set +o xtrace postgres-operator-778d48f559-2czmntrue + [[ -n '' ]] + desc 'create first cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first cluster ----------------------------------------------------------------------------------- + cluster=some-name + replica_quantity=3 + spinup_pgcluster some-name /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=scaling-8587 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets some-name scaling-8587 + local cluster=some-name + local namespace=scaling-8587 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name some-name-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n scaling-8587 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.PxMy7gA5U1 ++ mktemp + local LAST_ERR=/tmp/tmp.Xm0XcQ082j + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n scaling-8587 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PxMy7gA5U1 secret/some-name-users-secret created + cat /tmp/tmp.Xm0XcQ082j + rm /tmp/tmp.PxMy7gA5U1 /tmp/tmp.Xm0XcQ082j + return 0 + yq w - metadata.name some-name-pmm-secret + kubectl_bin -n scaling-8587 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.Mqyv8mkOwU ++ mktemp + local LAST_ERR=/tmp/tmp.MlLQu6lKuT + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n scaling-8587 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Mqyv8mkOwU secret/some-name-pmm-secret created + cat /tmp/tmp.MlLQu6lKuT + rm /tmp/tmp.Mqyv8mkOwU /tmp/tmp.MlLQu6lKuT + return 0 + kubectl_bin -n scaling-8587 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.5OrX3zE4KP ++ mktemp + local LAST_ERR=/tmp/tmp.MHKcjLtC7d + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n scaling-8587 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.5OrX3zE4KP deployment.apps/pg-client created + cat /tmp/tmp.MHKcjLtC7d + rm /tmp/tmp.5OrX3zE4KP /tmp/tmp.MHKcjLtC7d + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml some-name false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local name=some-name + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml metadata.annotations.current-primary some-name + yq w - metadata.labels.deployment-name some-name + yq w - metadata.labels.name some-name + yq w - metadata.name some-name + yq w - spec.database some-name + yq w - metadata.labels.pg-cluster some-name + yq w - spec.clustername some-name + yq w - spec.name some-name + yq w - spec.keepData false + yq w - spec.keepBackups false + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.user some-name + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.secretsName some-name-users-secret + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - metadata.labels.crunchy-pgha-scope some-name + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.WATPD0tkZe/cr.yaml -n scaling-8587 ++ mktemp + local LAST_OUT=/tmp/tmp.Me2KnSAXrZ ++ mktemp + local LAST_ERR=/tmp/tmp.OpidqfB8Jq + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.WATPD0tkZe/cr.yaml -n scaling-8587 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Me2KnSAXrZ perconapgcluster.pg.percona.com/some-name created + cat /tmp/tmp.OpidqfB8Jq + rm /tmp/tmp.Me2KnSAXrZ /tmp/tmp.OpidqfB8Jq + return 0 + wait_deployment some-name-backrest-shared-repo + local name=some-name-backrest-shared-repo + local target_namespace=scaling-8587 + sleep 10 + set +o xtrace some-name-backrest-shared-repo.... + wait_deployment some-name + local name=some-name + local target_namespace=scaling-8587 + sleep 10 + set +o xtrace some-name....... + [[ false == \g\c\s\+* ]] + wait_job_completion some-name-stanza-create false + local job=some-name-stanza-create + local mandatory=false + local ns=scaling-8587 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' some-name-stanza-create+ retry=0 ++ kubectl_bin -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-some-name + local job=backrest-backup-some-name + local mandatory=true + local ns=scaling-8587 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n scaling-8587 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' backrest-backup-some-name+ retry=0 ++ kubectl_bin -n scaling-8587 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency some-name + cluster_name=some-name + wait_cluster_status some-name 'pgcluster Initialized' + local cluster_name=some-name + local 'target_status=pgcluster Initialized' + local target_namespace=scaling-8587 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nqcuPFllgo +++ mktemp ++ local LAST_ERR=/tmp/tmp.z4n6XvbZY3 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.nqcuPFllgo ++ cat /tmp/tmp.z4n6XvbZY3 ++ rm /tmp/tmp.nqcuPFllgo /tmp/tmp.z4n6XvbZY3 ++ return 0 + status_query='kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.u2mpP4iavU ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fCa0zPjDCX +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.u2mpP4iavU +++ cat /tmp/tmp.fCa0zPjDCX +++ rm /tmp/tmp.u2mpP4iavU /tmp/tmp.fCa0zPjDCX +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=scaling-8587 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.f4sKec2ZCd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.YnyULcsRJt +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.f4sKec2ZCd +++ cat /tmp/tmp.YnyULcsRJt +++ rm /tmp/tmp.f4sKec2ZCd /tmp/tmp.YnyULcsRJt +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.G8GoESfpmR ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Dli6hnSSaF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.G8GoESfpmR +++ cat /tmp/tmp.Dli6hnSSaF +++ rm /tmp/tmp.G8GoESfpmR /tmp/tmp.Dli6hnSSaF +++ return 0 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.GIxFdU5gHH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BxiNuq0k7D +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.GIxFdU5gHH +++ cat /tmp/tmp.BxiNuq0k7D +++ rm /tmp/tmp.GIxFdU5gHH /tmp/tmp.BxiNuq0k7D +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.VacopsOi7I +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.4fkLfDbDo3 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.VacopsOi7I ++++ cat /tmp/tmp.4fkLfDbDo3 ++++ rm /tmp/tmp.VacopsOi7I /tmp/tmp.4fkLfDbDo3 ++++ return 0 +++ kubectl_bin -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.sOWPDmikK3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fqRNd54eJD +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.sOWPDmikK3 +++ cat /tmp/tmp.fqRNd54eJD +++ rm /tmp/tmp.sOWPDmikK3 /tmp/tmp.fqRNd54eJD +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@some-name.scaling-8587 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@some-name.scaling-8587 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I9ps9h7LXH +++ mktemp ++ local LAST_ERR=/tmp/tmp.6PM1UuGksY ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.I9ps9h7LXH ++ cat /tmp/tmp.6PM1UuGksY ++ rm /tmp/tmp.I9ps9h7LXH /tmp/tmp.6PM1UuGksY ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.bLgjXnEdGu ++ mktemp + local LAST_ERR=/tmp/tmp.kUWMIpyU1S + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.bLgjXnEdGu + cat /tmp/tmp.kUWMIpyU1S + rm /tmp/tmp.bLgjXnEdGu /tmp/tmp.kUWMIpyU1S + return 0 ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=scaling-8587 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6l9mcQODiz ++++ mktemp +++ local LAST_ERR=/tmp/tmp.WxVjPndLPj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.6l9mcQODiz +++ cat /tmp/tmp.WxVjPndLPj +++ rm /tmp/tmp.6l9mcQODiz /tmp/tmp.WxVjPndLPj +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.RJxHHHd2ML ++++ mktemp +++ local LAST_ERR=/tmp/tmp.QNUziUGXXp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.RJxHHHd2ML +++ cat /tmp/tmp.QNUziUGXXp +++ rm /tmp/tmp.RJxHHHd2ML /tmp/tmp.QNUziUGXXp +++ return 0 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.jYzDooPdZS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.rYS9KNlugu +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.jYzDooPdZS +++ cat /tmp/tmp.rYS9KNlugu +++ rm /tmp/tmp.jYzDooPdZS /tmp/tmp.rYS9KNlugu +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.OKrzajgmje +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.hGNOnGeiNb ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.OKrzajgmje ++++ cat /tmp/tmp.hGNOnGeiNb ++++ rm /tmp/tmp.OKrzajgmje /tmp/tmp.hGNOnGeiNb ++++ return 0 +++ kubectl_bin -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tUGnkEGLC1 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5e5ux2aee3 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tUGnkEGLC1 +++ cat /tmp/tmp.5e5ux2aee3 +++ rm /tmp/tmp.tUGnkEGLC1 /tmp/tmp.5e5ux2aee3 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@some-name.scaling-8587 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@some-name.scaling-8587 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6uqBS15zfM +++ mktemp ++ local LAST_ERR=/tmp/tmp.uqa8dEKcfQ ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.6uqBS15zfM ++ cat /tmp/tmp.uqa8dEKcfQ ++ rm /tmp/tmp.6uqBS15zfM /tmp/tmp.uqa8dEKcfQ ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.PHMb4bMpPN ++ mktemp + local LAST_ERR=/tmp/tmp.SVNWbXzOxg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PHMb4bMpPN + cat /tmp/tmp.SVNWbXzOxg + rm /tmp/tmp.PHMb4bMpPN /tmp/tmp.SVNWbXzOxg + return 0 ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=scaling-8587 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.kPoDiTlhtN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NUMzDAeQKj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.kPoDiTlhtN +++ cat /tmp/tmp.NUMzDAeQKj +++ rm /tmp/tmp.kPoDiTlhtN /tmp/tmp.NUMzDAeQKj +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tq5wcgqIWD ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wDVgJ7Wr2E +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tq5wcgqIWD +++ cat /tmp/tmp.wDVgJ7Wr2E +++ rm /tmp/tmp.tq5wcgqIWD /tmp/tmp.wDVgJ7Wr2E +++ return 0 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pCyFSAEgai ++++ mktemp +++ local LAST_ERR=/tmp/tmp.R7qwrjwfsZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.pCyFSAEgai +++ cat /tmp/tmp.R7qwrjwfsZ +++ rm /tmp/tmp.pCyFSAEgai /tmp/tmp.R7qwrjwfsZ +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.GXwV4nYuRs +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.SpPHBmOvID ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.GXwV4nYuRs ++++ cat /tmp/tmp.SpPHBmOvID ++++ rm /tmp/tmp.GXwV4nYuRs /tmp/tmp.SpPHBmOvID ++++ return 0 +++ kubectl_bin -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.NEamkspojc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XMCNLqKypd +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.NEamkspojc +++ cat /tmp/tmp.XMCNLqKypd +++ rm /tmp/tmp.NEamkspojc /tmp/tmp.XMCNLqKypd +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@some-name.scaling-8587 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@some-name.scaling-8587 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.45aXVYJEuV +++ mktemp ++ local LAST_ERR=/tmp/tmp.wUUFKxwdwv ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.45aXVYJEuV ++ cat /tmp/tmp.wUUFKxwdwv ++ rm /tmp/tmp.45aXVYJEuV /tmp/tmp.wUUFKxwdwv ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.TSsORqEUlg ++ mktemp + local LAST_ERR=/tmp/tmp.GfXuXe1VZC + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.TSsORqEUlg + cat /tmp/tmp.GfXuXe1VZC + rm /tmp/tmp.TSsORqEUlg /tmp/tmp.GfXuXe1VZC + return 0 + sleep 10 + desc 'scaling up' + set +o xtrace ----------------------------------------------------------------------------------- scaling up ----------------------------------------------------------------------------------- + kubectl_bin patch perconapgcluster/some-name --type json -p '[{"op": "replace", "path": "/spec/pgReplicas/hotStandby/size", "value": 3}]' ++ mktemp + local LAST_OUT=/tmp/tmp.q9q2k5MxjA ++ mktemp + local LAST_ERR=/tmp/tmp.9cbAqy70Xa + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl patch perconapgcluster/some-name --type json -p '[{"op": "replace", "path": "/spec/pgReplicas/hotStandby/size", "value": 3}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.q9q2k5MxjA perconapgcluster.pg.percona.com/some-name patched + cat /tmp/tmp.9cbAqy70Xa + rm /tmp/tmp.q9q2k5MxjA /tmp/tmp.9cbAqy70Xa + return 0 + kubectl_bin wait --for=condition=Ready --selector=pg-cluster=some-name,role=replica pods --timeout=120s ++ mktemp + local LAST_OUT=/tmp/tmp.szL4Qg9QZ8 ++ mktemp + local LAST_ERR=/tmp/tmp.dJBNB0m31B + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl wait --for=condition=Ready --selector=pg-cluster=some-name,role=replica pods --timeout=120s + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.szL4Qg9QZ8 pod/some-name-repl1-5489f574fb-pdjw8 condition met pod/some-name-repl2-759b4f7fdd-kttg9 condition met + cat /tmp/tmp.dJBNB0m31B + rm /tmp/tmp.szL4Qg9QZ8 /tmp/tmp.dJBNB0m31B + return 0 ++ seq 1 3 + for i in '$(seq 1 ${replica_quantity})' + compare_kubectl deployment/some-name-repl1 + local resource=deployment/some-name-repl1 + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/deployment_some-name-repl1.yml + local new_result=/tmp/tmp.WATPD0tkZe/deployment_some-name-repl1.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-repl1 + yq d - metadata.managedFields + yq d - '**.namespace' + yq d - '**.uid' + yq d - metadata.selfLink + yq d - metadata.deletionTimestamp + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - metadata.resourceVersion ++ mktemp + yq d - '**."batch.kubernetes.io/controller-uid"' + local LAST_OUT=/tmp/tmp.Go55G1Ucu6 + yq d - '**.creationTimestamp' ++ mktemp + local LAST_ERR=/tmp/tmp.Y7pmk7n9bH + local exit_status=0 + local timeout=4 + set +o errexit + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**.finalizers' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**.image' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**.storageClassName' + yq d - '**.clusterIP' + yq d - spec.volumeMode + yq d - '**.volumeName' + yq d - spec.nodeName + yq d - '**.clusterIPs' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - '**.procMount' + yq d - '**.dataSource' + yq d - spec.podReplacementPolicy + yq d - '**."kubernetes.io/pvc-protection"' + yq d - spec.manualSelector + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**."percona.com/*"' + yq d - '**.ipFamilies' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' ++ seq 0 2 + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.imagePullSecrets' + yq d - '**.healthCheckNodePort' + yq d - '**.controller-uid' + yq d - '**.preemptionPolicy' + yq d - '**.enableServiceLinks' + yq d - '**.nodePort' + yq d - '**.internalTrafficPolicy' + yq d - status + yq d - '**.ipFamilyPolicy' + yq d - '**.(name==suffix)' + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-repl1 + yq d - '**.(name==NAMESPACE)' + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.(name==PODNAME)' + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - '**.creationTimestamp' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Go55G1Ucu6 + cat /tmp/tmp.Y7pmk7n9bH + rm /tmp/tmp.Go55G1Ucu6 /tmp/tmp.Y7pmk7n9bH + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.WATPD0tkZe/deployment_some-name-repl1.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/deployment_some-name-repl1.yml /tmp/tmp.WATPD0tkZe/deployment_some-name-repl1.yml + for i in '$(seq 1 ${replica_quantity})' + compare_kubectl deployment/some-name-repl2 + local resource=deployment/some-name-repl2 + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/deployment_some-name-repl2.yml + local new_result=/tmp/tmp.WATPD0tkZe/deployment_some-name-repl2.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-repl2 + yq d - '**.creationTimestamp' ++ mktemp + yq d - '**.uid' + local LAST_OUT=/tmp/tmp.vb2jNc7mJu + yq d - metadata.resourceVersion + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - metadata.deletionTimestamp + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - metadata.selfLink ++ mktemp + yq d - '**.storageClassName' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**.finalizers' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**.procMount' + local LAST_ERR=/tmp/tmp.zw465W1vpw + local exit_status=0 + local timeout=4 + set +o errexit + yq d - '**.dataSource' + yq d - spec.nodeName + yq d - spec.manualSelector + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - '**.volumeName' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**.creationTimestamp' + yq d - spec.volumeMode + yq d - '**.image' + yq d - spec.podReplacementPolicy + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**.(name==NAMESPACE)' + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**."percona.com/*"' ++ seq 0 2 + for i in '$(seq 0 2)' + yq d - '**.(volumeMode==Filesystem).volumeMode' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-repl2 + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.healthCheckNodePort' + yq d - '**.(name==PODNAME)' + yq d - '**.nodePort' + yq d - '**.imagePullSecrets' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.clusterIPs' + yq d - '**.enableServiceLinks' + yq d - status + yq d - '**.(name==suffix)' + yq d - '**.clusterIP' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.ipFamilyPolicy' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.controller-uid' + yq d - '**.preemptionPolicy' + yq d - '**.ipFamilies' + yq d - '**.internalTrafficPolicy' + yq d - '**.namespace' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - metadata.managedFields + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.vb2jNc7mJu + cat /tmp/tmp.zw465W1vpw + rm /tmp/tmp.vb2jNc7mJu /tmp/tmp.zw465W1vpw + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.WATPD0tkZe/deployment_some-name-repl2.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/deployment_some-name-repl2.yml /tmp/tmp.WATPD0tkZe/deployment_some-name-repl2.yml + for i in '$(seq 1 ${replica_quantity})' + compare_kubectl deployment/some-name-repl3 + local resource=deployment/some-name-repl3 + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/deployment_some-name-repl3.yml + local new_result=/tmp/tmp.WATPD0tkZe/deployment_some-name-repl3.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-repl3 + yq d - metadata.managedFields + yq d - '**.creationTimestamp' + yq d - '**.namespace' + yq d - metadata.selfLink + yq d - metadata.deletionTimestamp + yq d - '**.uid' + yq d - metadata.resourceVersion + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - '**.creationTimestamp' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**.image' + yq d - '**.clusterIP' + yq d - '**.procMount' + yq d - '**.finalizers' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**.storageClassName' + yq d - spec.podReplacementPolicy + yq d - '**.clusterIPs' + yq d - spec.volumeMode + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - '**.dataSource' + yq d - spec.nodeName + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**.volumeName' + yq d - spec.manualSelector + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**."percona.com/*"' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.ipFamilyPolicy' + yq d - '**.healthCheckNodePort' + yq d - '**.controller-uid' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**.nodePort' + yq d - '**.imagePullSecrets' + yq d - '**.internalTrafficPolicy' + yq d - '**.enableServiceLinks' + yq d - status + yq d - '**.(name==suffix)' + yq d - '**.(name==NAMESPACE)' + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.(name==PODNAME)' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.preemptionPolicy' + yq d - '**.ipFamilies' ++ mktemp + local LAST_OUT=/tmp/tmp.UnptWOrudF ++ mktemp + local LAST_ERR=/tmp/tmp.WJZ3TFUBOK + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-repl3 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.UnptWOrudF + cat /tmp/tmp.WJZ3TFUBOK + rm /tmp/tmp.UnptWOrudF /tmp/tmp.WJZ3TFUBOK + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.WATPD0tkZe/deployment_some-name-repl3.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/deployment_some-name-repl3.yml /tmp/tmp.WATPD0tkZe/deployment_some-name-repl3.yml + sleep 10 ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=scaling-8587 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qEXJxwba6B ++++ mktemp +++ local LAST_ERR=/tmp/tmp.V3pwdLQXwy +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qEXJxwba6B +++ cat /tmp/tmp.V3pwdLQXwy +++ rm /tmp/tmp.qEXJxwba6B /tmp/tmp.V3pwdLQXwy +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dqZjSBZdyo ++++ mktemp +++ local LAST_ERR=/tmp/tmp.JwDgnXUlHY +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.dqZjSBZdyo +++ cat /tmp/tmp.JwDgnXUlHY +++ rm /tmp/tmp.dqZjSBZdyo /tmp/tmp.JwDgnXUlHY +++ return 0 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.sZuOpjDLxn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.om8CGki3JJ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.sZuOpjDLxn +++ cat /tmp/tmp.om8CGki3JJ +++ rm /tmp/tmp.sZuOpjDLxn /tmp/tmp.om8CGki3JJ +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.gJT9xE8jD7 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.EcOB8Lz0kQ ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.gJT9xE8jD7 ++++ cat /tmp/tmp.EcOB8Lz0kQ ++++ rm /tmp/tmp.gJT9xE8jD7 /tmp/tmp.EcOB8Lz0kQ ++++ return 0 +++ kubectl_bin -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rNTnyaR6l8 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.aIFxrnEuUF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rNTnyaR6l8 +++ cat /tmp/tmp.aIFxrnEuUF +++ rm /tmp/tmp.rNTnyaR6l8 /tmp/tmp.aIFxrnEuUF +++ return 0 ++ echo postgres_password + run_psql 'SELECT usename,application_name,client_addr,state from pg_stat_replication' postgres:postgres_password@some-name.scaling-8587 + local 'command=SELECT usename,application_name,client_addr,state from pg_stat_replication' + local uri=postgres:postgres_password@some-name.scaling-8587 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5enjQej2gr +++ mktemp ++ local LAST_ERR=/tmp/tmp.h8BQWv28gg ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.5enjQej2gr ++ cat /tmp/tmp.h8BQWv28gg ++ rm /tmp/tmp.5enjQej2gr /tmp/tmp.h8BQWv28gg ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''SELECT usename,application_name,client_addr,state from pg_stat_replication\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.2Yp3Sw9KOA ++ mktemp + local LAST_ERR=/tmp/tmp.YKhqtGu2BH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''SELECT usename,application_name,client_addr,state from pg_stat_replication\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.2Yp3Sw9KOA + cat /tmp/tmp.YKhqtGu2BH + rm /tmp/tmp.2Yp3Sw9KOA /tmp/tmp.YKhqtGu2BH + return 0 ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=scaling-8587 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DM2GFgzV0L ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Su6GXSB0vZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.DM2GFgzV0L +++ cat /tmp/tmp.Su6GXSB0vZ +++ rm /tmp/tmp.DM2GFgzV0L /tmp/tmp.Su6GXSB0vZ +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.k6LPQ43gu6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uhim6O5R3s +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.k6LPQ43gu6 +++ cat /tmp/tmp.uhim6O5R3s +++ rm /tmp/tmp.k6LPQ43gu6 /tmp/tmp.uhim6O5R3s +++ return 0 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.K25r6dd5nR ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gpdcF00d34 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.K25r6dd5nR +++ cat /tmp/tmp.gpdcF00d34 +++ rm /tmp/tmp.K25r6dd5nR /tmp/tmp.gpdcF00d34 +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.kjyrP4gIeP +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.8QvbN8MeyZ ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.kjyrP4gIeP ++++ cat /tmp/tmp.8QvbN8MeyZ ++++ rm /tmp/tmp.kjyrP4gIeP /tmp/tmp.8QvbN8MeyZ ++++ return 0 +++ kubectl_bin -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.bGAP0KYVVk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CEFlXSEx5a +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.bGAP0KYVVk +++ cat /tmp/tmp.CEFlXSEx5a +++ rm /tmp/tmp.bGAP0KYVVk /tmp/tmp.CEFlXSEx5a +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100501)' postgres:postgres_password@some-name.scaling-8587 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100501)' + local uri=postgres:postgres_password@some-name.scaling-8587 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SqGpKPKwcM +++ mktemp ++ local LAST_ERR=/tmp/tmp.QQn9WeiADc ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.SqGpKPKwcM ++ cat /tmp/tmp.QQn9WeiADc ++ rm /tmp/tmp.SqGpKPKwcM /tmp/tmp.QQn9WeiADc ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100501)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.ki7hRBIUaW ++ mktemp + local LAST_ERR=/tmp/tmp.KFCSqwvlLl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100501)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ki7hRBIUaW + cat /tmp/tmp.KFCSqwvlLl + rm /tmp/tmp.ki7hRBIUaW /tmp/tmp.KFCSqwvlLl + return 0 + sleep 15 ++ seq 1 3 + for i in '$(seq 1 ${replica_quantity})' ++ kubectl_bin get pods --selector=deployment-name=some-name-repl1 -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JKldlWSyza +++ mktemp ++ local LAST_ERR=/tmp/tmp.MtuQh4B96p ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=deployment-name=some-name-repl1 -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.JKldlWSyza ++ cat /tmp/tmp.MtuQh4B96p ++ rm /tmp/tmp.JKldlWSyza /tmp/tmp.MtuQh4B96p ++ return 0 + replica_pod_name=some-name-repl1-5489f574fb-pdjw8 ++ grep some-name-repl1-5489f574fb-pdjw8 /tmp/tmp.WATPD0tkZe/replicas.list ++ grep streaming + [[ -z primaryuser | some-name-repl1-5489f574fb-pdjw8 | 10.87.33.42 | streaming ]] ++ kubectl_bin get pods --selector=deployment-name=some-name-repl1 -o 'jsonpath={.items[0].status.podIP}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JCNTEU7UZC +++ mktemp ++ local LAST_ERR=/tmp/tmp.mBV0kepZ04 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=deployment-name=some-name-repl1 -o 'jsonpath={.items[0].status.podIP}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.JCNTEU7UZC ++ cat /tmp/tmp.mBV0kepZ04 ++ rm /tmp/tmp.JCNTEU7UZC /tmp/tmp.mBV0kepZ04 ++ return 0 + replica_pod_IP=10.87.33.42 + compare_psql_cmd select-2 '\c myapp \\\ SELECT * from myApp;' some-name:some-name_pass@10.87.33.42 '' ' ' + local command_id=select-2 + local 'command=\c myapp \\\ SELECT * from myApp;' + local uri=some-name:some-name_pass@10.87.33.42 + local postfix= + local 'uri_suffix= ' + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/select-2.sql + run_psql '\c myapp \\\ SELECT * from myApp;' some-name:some-name_pass@10.87.33.42 postgres ' ' + local 'command=\c myapp \\\ SELECT * from myApp;' + local uri=some-name:some-name_pass@10.87.33.42 + local driver=postgres + local 'suffix= ' ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.G5Zh0Eg9ZB +++ mktemp ++ local LAST_ERR=/tmp/tmp.hPyeRYJIn5 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.G5Zh0Eg9ZB ++ cat /tmp/tmp.hPyeRYJIn5 ++ rm /tmp/tmp.G5Zh0Eg9ZB /tmp/tmp.hPyeRYJIn5 ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://some-name:some-name_pass@10.87.33.42 ' ++ mktemp + local LAST_OUT=/tmp/tmp.nDPoS2p6LT ++ mktemp + local LAST_ERR=/tmp/tmp.TgCacxIVj0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://some-name:some-name_pass@10.87.33.42 ' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nDPoS2p6LT + cat /tmp/tmp.TgCacxIVj0 + rm /tmp/tmp.nDPoS2p6LT /tmp/tmp.TgCacxIVj0 + return 0 + '[' '!' -s /tmp/tmp.WATPD0tkZe/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/select-2.sql /tmp/tmp.WATPD0tkZe/select-2.sql + for i in '$(seq 1 ${replica_quantity})' ++ kubectl_bin get pods --selector=deployment-name=some-name-repl2 -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k5hNjhxePO +++ mktemp ++ local LAST_ERR=/tmp/tmp.9cQWHWZZw2 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=deployment-name=some-name-repl2 -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.k5hNjhxePO ++ cat /tmp/tmp.9cQWHWZZw2 ++ rm /tmp/tmp.k5hNjhxePO /tmp/tmp.9cQWHWZZw2 ++ return 0 + replica_pod_name=some-name-repl2-759b4f7fdd-kttg9 ++ grep some-name-repl2-759b4f7fdd-kttg9 /tmp/tmp.WATPD0tkZe/replicas.list ++ grep streaming + [[ -z primaryuser | some-name-repl2-759b4f7fdd-kttg9 | 10.87.32.27 | streaming ]] ++ kubectl_bin get pods --selector=deployment-name=some-name-repl2 -o 'jsonpath={.items[0].status.podIP}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7cPxs5N7XA +++ mktemp ++ local LAST_ERR=/tmp/tmp.kjmMzWAa0Y ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=deployment-name=some-name-repl2 -o 'jsonpath={.items[0].status.podIP}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.7cPxs5N7XA ++ cat /tmp/tmp.kjmMzWAa0Y ++ rm /tmp/tmp.7cPxs5N7XA /tmp/tmp.kjmMzWAa0Y ++ return 0 + replica_pod_IP=10.87.32.27 + compare_psql_cmd select-2 '\c myapp \\\ SELECT * from myApp;' some-name:some-name_pass@10.87.32.27 '' ' ' + local command_id=select-2 + local 'command=\c myapp \\\ SELECT * from myApp;' + local uri=some-name:some-name_pass@10.87.32.27 + local postfix= + local 'uri_suffix= ' + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/select-2.sql + run_psql '\c myapp \\\ SELECT * from myApp;' some-name:some-name_pass@10.87.32.27 postgres ' ' + local 'command=\c myapp \\\ SELECT * from myApp;' + local uri=some-name:some-name_pass@10.87.32.27 + local driver=postgres + local 'suffix= ' ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hSkpDSzrUK +++ mktemp ++ local LAST_ERR=/tmp/tmp.uf3Il5C9XX ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.hSkpDSzrUK ++ cat /tmp/tmp.uf3Il5C9XX ++ rm /tmp/tmp.hSkpDSzrUK /tmp/tmp.uf3Il5C9XX ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://some-name:some-name_pass@10.87.32.27 ' ++ mktemp + local LAST_OUT=/tmp/tmp.7Uh6zBUZcX ++ mktemp + local LAST_ERR=/tmp/tmp.3qrNmf0Bb0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://some-name:some-name_pass@10.87.32.27 ' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.7Uh6zBUZcX + cat /tmp/tmp.3qrNmf0Bb0 + rm /tmp/tmp.7Uh6zBUZcX /tmp/tmp.3qrNmf0Bb0 + return 0 + '[' '!' -s /tmp/tmp.WATPD0tkZe/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/select-2.sql /tmp/tmp.WATPD0tkZe/select-2.sql + for i in '$(seq 1 ${replica_quantity})' ++ kubectl_bin get pods --selector=deployment-name=some-name-repl3 -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5BogeAszXw +++ mktemp ++ local LAST_ERR=/tmp/tmp.DXCzlU7DWB ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=deployment-name=some-name-repl3 -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.5BogeAszXw ++ cat /tmp/tmp.DXCzlU7DWB ++ rm /tmp/tmp.5BogeAszXw /tmp/tmp.DXCzlU7DWB ++ return 0 + replica_pod_name=some-name-repl3-5fc65cdbb8-rtlq2 ++ grep some-name-repl3-5fc65cdbb8-rtlq2 /tmp/tmp.WATPD0tkZe/replicas.list ++ grep streaming + [[ -z primaryuser | some-name-repl3-5fc65cdbb8-rtlq2 | 10.87.34.54 | streaming ]] ++ kubectl_bin get pods --selector=deployment-name=some-name-repl3 -o 'jsonpath={.items[0].status.podIP}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.931k0VrVHf +++ mktemp ++ local LAST_ERR=/tmp/tmp.fruTuQdHst ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=deployment-name=some-name-repl3 -o 'jsonpath={.items[0].status.podIP}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.931k0VrVHf ++ cat /tmp/tmp.fruTuQdHst ++ rm /tmp/tmp.931k0VrVHf /tmp/tmp.fruTuQdHst ++ return 0 + replica_pod_IP=10.87.34.54 + compare_psql_cmd select-2 '\c myapp \\\ SELECT * from myApp;' some-name:some-name_pass@10.87.34.54 '' ' ' + local command_id=select-2 + local 'command=\c myapp \\\ SELECT * from myApp;' + local uri=some-name:some-name_pass@10.87.34.54 + local postfix= + local 'uri_suffix= ' + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/select-2.sql + run_psql '\c myapp \\\ SELECT * from myApp;' some-name:some-name_pass@10.87.34.54 postgres ' ' + local 'command=\c myapp \\\ SELECT * from myApp;' + local uri=some-name:some-name_pass@10.87.34.54 + local driver=postgres + local 'suffix= ' ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ldqxn6vDIn +++ mktemp ++ local LAST_ERR=/tmp/tmp.KHAWY2W0MC ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.Ldqxn6vDIn ++ cat /tmp/tmp.KHAWY2W0MC ++ rm /tmp/tmp.Ldqxn6vDIn /tmp/tmp.KHAWY2W0MC ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://some-name:some-name_pass@10.87.34.54 ' ++ mktemp + local LAST_OUT=/tmp/tmp.kfJfOg1AHE ++ mktemp + local LAST_ERR=/tmp/tmp.Xs5ZIdiGvT + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://some-name:some-name_pass@10.87.34.54 ' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.kfJfOg1AHE + cat /tmp/tmp.Xs5ZIdiGvT + rm /tmp/tmp.kfJfOg1AHE /tmp/tmp.Xs5ZIdiGvT + return 0 + '[' '!' -s /tmp/tmp.WATPD0tkZe/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/scaling/compare/select-2.sql /tmp/tmp.WATPD0tkZe/select-2.sql + desc 'Scaling down' + set +o xtrace ----------------------------------------------------------------------------------- Scaling down ----------------------------------------------------------------------------------- + kubectl_bin patch perconapgcluster/some-name --type json -p '[{"op": "replace", "path": "/spec/pgReplicas/hotStandby/size", "value": 0}]' ++ mktemp + local LAST_OUT=/tmp/tmp.8YroHEJa1e ++ mktemp + local LAST_ERR=/tmp/tmp.NirpEuNjD3 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl patch perconapgcluster/some-name --type json -p '[{"op": "replace", "path": "/spec/pgReplicas/hotStandby/size", "value": 0}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.8YroHEJa1e perconapgcluster.pg.percona.com/some-name patched + cat /tmp/tmp.NirpEuNjD3 + rm /tmp/tmp.8YroHEJa1e /tmp/tmp.NirpEuNjD3 + return 0 + kubectl_bin wait --for=delete --selector=pg-cluster=some-name,role=replica pods --timeout=180s ++ mktemp + local LAST_OUT=/tmp/tmp.iCYNz4vWcx ++ mktemp + local LAST_ERR=/tmp/tmp.xIahudo51K + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl wait --for=delete --selector=pg-cluster=some-name,role=replica pods --timeout=180s + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.iCYNz4vWcx pod/some-name-repl1-5489f574fb-pdjw8 condition met pod/some-name-repl2-759b4f7fdd-kttg9 condition met pod/some-name-repl3-5fc65cdbb8-rtlq2 condition met + cat /tmp/tmp.xIahudo51K + rm /tmp/tmp.iCYNz4vWcx /tmp/tmp.xIahudo51K + return 0 ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=scaling-8587 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mgsYWC6RM9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.4Xvo33FYUR +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.mgsYWC6RM9 +++ cat /tmp/tmp.4Xvo33FYUR +++ rm /tmp/tmp.mgsYWC6RM9 /tmp/tmp.4Xvo33FYUR +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lrMHTBaJgM ++++ mktemp +++ local LAST_ERR=/tmp/tmp.MBsBDout34 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.lrMHTBaJgM +++ cat /tmp/tmp.MBsBDout34 +++ rm /tmp/tmp.lrMHTBaJgM /tmp/tmp.MBsBDout34 +++ return 0 +++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.85GvUjdKPY ++++ mktemp +++ local LAST_ERR=/tmp/tmp.I3bHAGF84e +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.85GvUjdKPY +++ cat /tmp/tmp.I3bHAGF84e +++ rm /tmp/tmp.85GvUjdKPY /tmp/tmp.I3bHAGF84e +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.rTiwGcfHXf +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.kCRG4HJOce ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n scaling-8587 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.rTiwGcfHXf ++++ cat /tmp/tmp.kCRG4HJOce ++++ rm /tmp/tmp.rTiwGcfHXf /tmp/tmp.kCRG4HJOce ++++ return 0 +++ kubectl_bin -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.F2pHQXs3bn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.MiLxfza3I8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n scaling-8587 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.F2pHQXs3bn +++ cat /tmp/tmp.MiLxfza3I8 +++ rm /tmp/tmp.F2pHQXs3bn /tmp/tmp.MiLxfza3I8 +++ return 0 ++ echo postgres_password + run_psql 'SELECT usename,application_name,client_addr,state from pg_stat_replication' postgres:postgres_password@some-name.scaling-8587 + local 'command=SELECT usename,application_name,client_addr,state from pg_stat_replication' + local uri=postgres:postgres_password@some-name.scaling-8587 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4ufntTvH18 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Pjd4Tar4ZN ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n scaling-8587 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.4ufntTvH18 ++ cat /tmp/tmp.Pjd4Tar4ZN ++ rm /tmp/tmp.4ufntTvH18 /tmp/tmp.Pjd4Tar4ZN ++ return 0 + local client_container=pg-client-76df9df876-rqt5p + kubectl_bin exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''SELECT usename,application_name,client_addr,state from pg_stat_replication\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.LNT5bmJhDk ++ mktemp + local LAST_ERR=/tmp/tmp.IoTbqWf4uJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-rqt5p -n scaling-8587 -- bash -c 'printf '\''SELECT usename,application_name,client_addr,state from pg_stat_replication\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.scaling-8587.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.LNT5bmJhDk + cat /tmp/tmp.IoTbqWf4uJ + rm /tmp/tmp.LNT5bmJhDk /tmp/tmp.IoTbqWf4uJ + return 0 ++ kubectl_bin get pods --selector=name=some-name-replica -o 'go-template={{len .items}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OwkFwsQxrk +++ mktemp ++ local LAST_ERR=/tmp/tmp.VfJOk1AAud ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=some-name-replica -o 'go-template={{len .items}}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.OwkFwsQxrk ++ cat /tmp/tmp.VfJOk1AAud ++ rm /tmp/tmp.OwkFwsQxrk /tmp/tmp.VfJOk1AAud ++ return 0 + [[ 0 -gt 0 ]] + destroy scaling-8587 + grep -v level=debug + grep -v 'Getting tasks for pod' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' + grep -v 'get backup status: Job.batch' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + grep -v 'the object has been modified' + grep -v 'Getting pods from source' + grep -v level=info +++ mktemp + sort -u + tee /tmp/tmp.WATPD0tkZe/operator.log ++ local LAST_OUT=/tmp/tmp.gw4UnGsJRK +++ mktemp ++ local LAST_ERR=/tmp/tmp.i3Fc20defX ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.gw4UnGsJRK ++ cat /tmp/tmp.i3Fc20defX ++ rm /tmp/tmp.gw4UnGsJRK /tmp/tmp.i3Fc20defX ++ return 0 + kubectl_bin logs postgres-operator-778d48f559-2czmn -c operator ++ mktemp + local LAST_OUT=/tmp/tmp.falyw79CFS ++ mktemp + local LAST_ERR=/tmp/tmp.V72OkWckjq + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl logs postgres-operator-778d48f559-2czmn -c operator + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.falyw79CFS + cat /tmp/tmp.V72OkWckjq + rm /tmp/tmp.falyw79CFS /tmp/tmp.V72OkWckjq + return 0 ] { } }, ] } }, { } }, ] { } }, { } ] } { } }, ] } ] { } }, }, { ] { } }, ] ] ], { } }, }, { }, { },{ }] }], ] { } }, , ], { } }, }], { } }, }], } ], } }, } }, } }, { } }, }, 2024/05/24 19:44:47 INF 1 (localhost:4150) connecting to nsqd 2024/05/24 19:45:40 INF 2 (localhost:4150) connecting to nsqd 2024/05/24 19:45:41 INF 3 (localhost:4150) connecting to nsqd 2024/05/24 19:45:47 INF 4 (localhost:4150) connecting to nsqd 2024/05/24 19:46:13 INF 5 (localhost:4150) connecting to nsqd 2024/05/24 19:46:14 INF 6 (localhost:4150) connecting to nsqd 2024/05/24 19:46:14 INF 7 (localhost:4150) connecting to nsqd 2024/05/24 19:46:59 INF 8 (localhost:4150) connecting to nsqd 2024/05/24 19:49:08 INF 9 (localhost:4150) connecting to nsqd 2024/05/24 19:49:13 INF 10 (localhost:4150) connecting to nsqd 2024/05/24 19:49:18 INF 11 (localhost:4150) connecting to nsqd "accessModes": [ "affinity": "affinity": { "affinity": { "podAntiAffinity": { "allowPrivilegeEscalation": false, "annotations": {"keep-backups":"false","keep-data":"false"}, "apiVersion": "apps/v1", "apiVersion": "batch/v1", "apiVersion": "v1", "apiVersion": "v1", "apiVersion": "v1", "backrest-command": "backup", "backrest-command": "stanza-create", "claimName": "some-name-pgbr-repo" "command": [ "configMap": { "configMap": { "configMapKeyRef": { "containerName": "database", "containerPort": 2022, "containerPort": 5432, "containerPort": 8009, "containers": [ "containers": [{ "cpu": "1", "cpu": "200m", "cpu": "500m", "crunchydata" "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "defaultMode": 288 "defaultMode": 288, "defaultMode": 288, "defaultMode": 420, "deployment-name": "some-name" "deployment-name": "some-name","pgouser": "admin","crunchy-pgha-scope": "some-name","pgo-version": "1.7.0","name": "some-name","pg-cluster": "some-name" "deployment-name": "some-name-repl1" "deployment-name": "some-name-repl1","pgo-version": "1.7.0","name": "some-name-replica","pg-cluster": "some-name" "deployment-name": "some-name-repl2" "deployment-name": "some-name-repl3" "divisor": "1m", "dnsPolicy": "ClusterFirst" "downwardAPI": { "emptyDir": { "emptyDir": { "emptyDir": { "emptyDir": { "env": [ "env": [{ "exec": { "fieldPath": "metadata.annotations" "fieldPath": "metadata.labels" "fieldPath": "metadata.labels['crunchy-pgha-scope']" "fieldPath": "metadata.namespace" "fieldRef": { "fieldRef": { "fsGroup": 2, I0524 19:44:12.109889 1 shared_informer.go:240] Waiting for caches to sync for namespace I0524 19:44:12.210840 1 shared_informer.go:247] Caches are synced for namespace I0524 19:44:12.362050 1 shared_informer.go:240] Waiting for caches to sync for scaling-8587 I0524 19:44:12.463203 1 shared_informer.go:247] Caches are synced for scaling-8587 "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbouncer", "image": "perconalab/percona-postgresql-operator:main-ppg14-postgres-ha", "imagePullPolicy": "IfNotPresent" "imagePullPolicy": "IfNotPresent", "initialDelaySeconds": 15 "initialDelaySeconds": 30, "items": [ "items": [ "items": [ "key": "config", "key": "config", "key": "crunchy-pgbouncer", "key": "init" "key": "password" "key": "pg-cluster", "key": "pgo-backrest", "key": "pgo-backrest-job", "key": "pgo-pg-database", "key": "pg-pod-anti-affinity", "key": "tls.crt", "key": "tls.key", "key": "vendor", "kind": "Deployment", "kind": "Job", "kind": "PersistentVolumeClaim", "kind": "Service", "labels": { "labels": { "labels": { "labelSelector": { "livenessProbe": { "matchExpressions": [ "matchLabels": { "matchLabels": { "maxSurge": 1 "maxUnavailable": 1, "medium": "Memory", "medium": "Memory" "medium": "Memory", "medium": "Memory", "medium": "Memory", "memory": "128Mi" "memory": "256Mi" "memory": "48Mi" "metadata": { "metadata": { "metadata": { "mountPath": "/backrestrepo", "mountPath": "/dev/shm", "mountPath": "/etc/pgbackrest" "mountPath": "/etc/pgbackrest/conf.d", "mountPath": "/etc/podinfo", "mountPath": "/etc/ssh", "mountPath": "/etc/ssh", "mountPath": "/home/pgbackrest" "mountPath": "/pgconf", "mountPath": "/pgconf/", "mountPath": "/pgconf/pgreplicator", "mountPath": "/pgconf/pgsuper", "mountPath": "/pgconf/pguser", "mountPath": "/pgconf/tls", "mountPath": "/pgconf/tls/pgbouncer", "mountPath": "/pgconf/tls-replication", "mountPath": "/pgdata", "mountPath": "/sshd", "mountPath": "/sshd", "mountPath": "/tmp", "mountPath": "/tmp", "mountPath": "/tmp" "name": "backrest", "name": "backrest-backup-some-name", "name": "backrestrepo", "name": "backrestrepo", "name": "BACKREST_SKIP_CREATE_STANZA", "name": "cluster1-replication-ssl-keypair", "name": "cluster1-ssl-ca" "name": "cluster1-ssl-ca" "name": "cluster1-ssl-keypair" "name": "cluster1-ssl-keypair" "name": "COMMAND", "name": "COMMAND_OPTS", "name": "database", "name": "dshm" "name": "dshm", "name": "ENABLE_SSHD", "name": "LD_PRELOAD", "name": "LD_PRELOAD", "name": "MODE", "name": "NAMESPACE", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_PASSWD", "name": "NSS_WRAPPER_PASSWD", "name": "patroni" "name": "PATRONI_KUBERNETES_LABELS", "name": "PATRONI_KUBERNETES_NAMESPACE", "name": "PATRONI_KUBERNETES_SCOPE_LABEL", "name": "PATRONI_LOG_LEVEL", "name": "PATRONI_POSTGRESQL_DATA_DIR", "name": "PATRONI_SCOPE", "name": "pgbackrest" "name": "pgbackrest-conf", "name": "pgbackrest-conf", "name": "pgbackrest-config" "name": "pgbackrest-config", "name": "PGBACKREST_DB_HOST", "name": "PGBACKREST_DB_PATH", "name": "PGBACKREST_DB_PATH", "name": "pgbackrest-home", "name": "pgbackrest-home", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_REPO1_HOST", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_STANZA", "name": "PGBACKREST_STANZA", "name": "pgbouncer" "name": "pgbouncer", "name": "pgbouncer-conf", "name": "pgbouncer-conf", "name": "pgconf-volume" "name": "pgconf-volume", "name": "pgdata", "name": "pgdata", "name": "PGHA_DATABASE", "name": "PGHA_INIT", "name": "PGHA_PASSWORD_TYPE", "name": "PGHA_PGBACKREST", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_S3_VERIFY_TLS", "name": "PGHA_PG_PORT", "name": "PGHA_REPLICA_REINIT_ON_START_FAIL", "name": "PGHA_STANDBY", "name": "PGHA_SYNC_REPLICATION", "name": "PGHA_TLS_ENABLED", "name": "PGHA_TLS_ONLY", "name": "PGHA_USER", "name": "PGHOST", "name": "PG_PASSWORD", "name": "PG_PRIMARY_SERVICE_NAME", "name": "PITR_TARGET", "name": "podinfo" "name": "podinfo", "name": "PODNAME", "name": "postgres" "name": "primary-volume" "name": "primary-volume", "name": "report", "name": "root-volume" "name": "root-volume", "name": "some-name", "name": "some-name", "name": "some-name-backrest-shared-repo", "name": "some-name-backrest-shared-repo" "name": "some-name-backrest-shared-repo", "name": "some-name-backrest-shared-repo", "name": "some-name-backrest-shared-repo", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer-cm" "name": "some-name-pgbouncer-secret", "name": "some-name-pgbouncer-secret", "name": "some-name-pgbr-repo", "name": "some-name-pgha-config", "name": "some-name-pgha-config", "name": "some-name-repl1", "name": "some-name-repl1", "name": "some-name-repl2", "name": "some-name-repl2", "name": "some-name-repl3", "name": "some-name-repl3", "name": "some-name-stanza-create", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "sshd", "name": "sshd", "name": "sshd", "name": "sshd", "name": "tls-pgbouncer" "name": "tls-pgbouncer", "name": "tls-replication" "name": "tls-replication", "name": "tls-server" "name": "tls-server", "name": "tmp" "name": "tmp" "name": "tmp", "name": "tmp", "name": "tmp", "name": "user-volume" "name": "user-volume", "nodePort": 0 "operator": "Exists" "operator": "In", "/opt/crunchy/bin/postgres-ha/health/pgha-liveness.sh" "/opt/crunchy/bin/postgres-ha/health/pgha-readiness.sh" "optional": true "path": "annotations" "path": "cpu_limit", "path": "cpu_request", "path": "labels" "path": "mem_limit", "path": "mem_request", "path": "ssh_config" "path": "ssh_config" "path": "tls-replication.crt" "path": "tls-replication.key" "periodSeconds": 15, "persistentVolumeClaim": { "persistentVolumeClaim":{"claimName":"some-name"} "persistentVolumeClaim":{"claimName":"some-name-repl1"} "persistentVolumeClaim":{"claimName":"some-name-repl2"} "persistentVolumeClaim":{"claimName":"some-name-repl3"} "pg-cluster": "some-name" "pg-cluster": "some-name", "pg-cluster": "some-name" "pg-cluster": "some-name", "pg-cluster": "some-name" "pg-cluster": "some-name", "pg-cluster": "some-name","deployment-name": "some-name-repl3","pgo-version": "1.7.0","name": "some-name-replica" "pgo-backrest-job": "true", "pgo-backrest-repo": "true" "pgo-backrest-repo": "true", "pgo-backrest-repo": "true" "pgo-backrest": "true", "pgo-pg-database": "true", "pgo-pg-database": "true", "pgo-version": "1.7.0","name": "some-name","pg-cluster": "some-name","deployment-name": "some-name","pgouser": "admin","crunchy-pgha-scope": "some-name" "pgo-version": "1.7.0","name": "some-name-replica","pg-cluster": "some-name","deployment-name": "some-name-repl1" "pgo-version": "1.7.0","name": "some-name-replica","pg-cluster": "some-name","deployment-name": "some-name-repl2" "pgo-version": "1.7.0","name": "some-name-replica","pg-cluster": "some-name","deployment-name": "some-name-repl2" "pgo-version": "1.7.0","name": "some-name-replica","pg-cluster": "some-name","deployment-name": "some-name-repl3" "pg-pod-anti-affinity": "preferred", "pg-pod-anti-affinity": "preferred", "podAffinityTerm": { "podAntiAffinity": { "podAntiAffinity": { "podAntiAffinity": { "port": 2022, "ports": [{ "ports": [{ "preferredDuringSchedulingIgnoredDuringExecution": [ "privileged": false, "projected": { "projected": { "projected": { "projected": { "sources": [] } "protocol": "TCP", "protocol": "TCP", "readinessProbe": { "readOnly": false "readOnly": false "readOnly": false "readOnlyRootFilesystem": true "readOnly": true "readOnly": true "readOnly": true "ReadWriteOnce" "replicas": 0, "replicas": 1, "replicas": 3, "requests": { "requests": { "resourceFieldRef": { "resource": "limits.cpu" "resource": "limits.memory" "resource": "requests.cpu" "resource": "requests.memory" "resources": { "resources": { "restartPolicy": "Always", "restartPolicy": "Never" "rollingUpdate": { "runAsNonRoot": true "secret": { "secret": { "secret": { "secret": { "secret": { "secretKeyRef": { "secretName": "some-name-backrest-repo-config" "secretName": "some-name-backrest-repo-config", "secretName": "some-name-backrest-repo-config" "secretName": "some-name-backrest-repo-config", "secretName": "some-name-postgres-secret" "secretName": "some-name-primaryuser-secret" "secretName": "some-name-some-name-secret" "securityContext": { "securityContext": { "securityContext": {}, "securityContext": {"fsGroup": 26,"supplementalGroups": [1001]}, "securityContext": {"runAsNonRoot":true,"fsGroup":26}, "selector": { "selector": { "serviceAccountName": "pgo-backrest", "serviceAccountName": "pgo-default", "serviceAccountName": "pgo-pg", "service-name": "some-name-backrest-shared-repo", "service-name": "some-name-backrest-shared-repo", "service-name": "some-name-pgbouncer", "service-name": "some-name-pgbouncer", "service-name": "some-name-pgbouncer", "sessionAffinity": "None" "sizeLimit": "128Ki" "sizeLimit": "16Mi" "sizeLimit": "1Mi" "sizeLimit": "2Mi" "sizeLimit": "64Mi" "sizeLimit": "64Mi" "some-name" "sources": [ "sources": [ "sources": [ "spec": { "spec": { "spec": { "storage": "1Gi" "strategy": { "targetPort": 2022, "template": { time="2024-05-24T19:44:11Z" level=warning msg="PGAdminStorage setting not set, will use PrimaryStorage setting" func="github.com/percona/percona-postgresql-operator/internal/config.(*PgoConfig).Validate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/config/pgoconfig.go:337" version=1.7.0 time="2024-05-24T19:44:51Z" level=error msg="get deployment some-name-repl1: deployments.apps \"some-name-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:44:56Z" level=error msg="get deployment some-name-repl2: deployments.apps \"some-name-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:45:40Z" level=error msg="Error completing workflow some-name-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:45:40Z" level=error msg="pgtasks.pg.percona.com \"some-name-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:46:15Z" level=warning msg="crv1 pgreplica some-name-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:46:15Z" level=warning msg="crv1 pgreplica some-name-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:46:59Z" level=warning msg="crv1 pgreplica some-name-repl3 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:47:12Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T19:47:12Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:47:12Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 "timeoutSeconds": 10 "topologyKey": "kubernetes.io/hostname" "type": "ClusterIP", "type": "RollingUpdate", "value": "" "value": "5432" "value": "5432" "value": "/backrestrepo/some-name-backrest-shared-repo" "value": "/backrestrepo/some-name-backrest-shared-repo" "value": "backup" "value": "crunchy-pgha-scope" "value": "db" "value": "db" "value": " --db-host=10.87.34.52 --db-path=/pgdata/some-name" "value": "false" "value": "false" "valueFrom": { "value": "INFO" "value": "pgbackrest" "value": "pgbackrest-repo" "value": "/pgdata/some-name" "value": "/pgdata/some-name" "value": "/pgdata/some-name-repl1" "value": "/pgdata/some-name-repl1" "value": "/pgdata/some-name-repl2" "value": "/pgdata/some-name-repl2" "value": "/pgdata/some-name-repl3" "value": "/pgdata/some-name-repl3" "value": "posix" "value": "posix" "value": "postgres" "values": [ "value": "some-name" "value": "some-name-backrest-shared-repo" "value": "some-name-backrest-shared-repo-6f8f947dbc-nl82c" "value": "stanza-create" "value": "/tmp" "value": "/tmp" "value": "/tmp/nss_wrapper/pgbackrest-repo/group" "value": "/tmp/nss_wrapper/pgbackrest-repo/passwd" "value": "/tmp/nss_wrapper/postgres/group" "value": "/tmp/nss_wrapper/postgres/passwd" "value": "true" "value": "true" "value": "--type=full --db-host=10.87.34.52 --db-path=/pgdata/some-name" "value": "/usr/lib64/libnss_wrapper.so" "value": "/usr/lib64/libnss_wrapper.so" "value": "{vendor: \"crunchydata\"}" "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "volumeMounts": [ "volumeMounts": [{ "volumes": [ "volumes": [{ "weight": 1, + destroy_operator + kubectl_bin delete -f /tmp/tmp.WATPD0tkZe/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.R8Y2Etvw1g ++ mktemp + local LAST_ERR=/tmp/tmp.7HkBpVw2Kj + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.WATPD0tkZe/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.WATPD0tkZe/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.WATPD0tkZe/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.R8Y2Etvw1g + cat /tmp/tmp.7HkBpVw2Kj Error from server (NotFound): error when deleting "/tmp/tmp.WATPD0tkZe/operator.yaml": serviceaccounts "pgo-deployer-sa" not found Error from server (NotFound): error when deleting "/tmp/tmp.WATPD0tkZe/operator.yaml": clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): error when deleting "/tmp/tmp.WATPD0tkZe/operator.yaml": configmaps "pgo-deployer-cm" not found Error from server (NotFound): error when deleting "/tmp/tmp.WATPD0tkZe/operator.yaml": clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-crb" not found Error from server (NotFound): error when deleting "/tmp/tmp.WATPD0tkZe/operator.yaml": jobs.batch "pgo-deploy" not found + rm /tmp/tmp.R8Y2Etvw1g /tmp/tmp.7HkBpVw2Kj + return 1 + true + yq w -d4 /tmp/tmp.WATPD0tkZe/operator.yaml 'spec.template.spec.containers[0].env[0].value' uninstall + kubectl_bin apply -f /tmp/tmp.WATPD0tkZe/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.Q2xfcZDrQe ++ mktemp + local LAST_ERR=/tmp/tmp.2Z9tsSGOPg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.WATPD0tkZe/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Q2xfcZDrQe serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb created job.batch/pgo-deploy created + cat /tmp/tmp.2Z9tsSGOPg + rm /tmp/tmp.Q2xfcZDrQe /tmp/tmp.2Z9tsSGOPg + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=scaling-8587 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 15 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 16 -ge 600 ']' ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n scaling-8587 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.WATPD0tkZe/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.dKqNKa1WF7 ++ mktemp + local LAST_ERR=/tmp/tmp.R1nBUMrE8a + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.WATPD0tkZe/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.dKqNKa1WF7 serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.R1nBUMrE8a + rm /tmp/tmp.dKqNKa1WF7 /tmp/tmp.R1nBUMrE8a + return 0 + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + awk '{print$1}' + xargs kubectl delete ns --grace-period=0 --force=true ++ mktemp + local LAST_OUT=/tmp/tmp.SWM6NI1PpY ++ mktemp + local LAST_ERR=/tmp/tmp.rRpX85WZcq + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.SWM6NI1PpY + cat /tmp/tmp.rRpX85WZcq + rm /tmp/tmp.SWM6NI1PpY /tmp/tmp.rRpX85WZcq + return 0 Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. namespace "scaling-8587" force deleted + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.LxaJKTPRJo ++ mktemp + local LAST_ERR=/tmp/tmp.6KVOpqkFIM + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.LxaJKTPRJo + cat /tmp/tmp.6KVOpqkFIM Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.LxaJKTPRJo /tmp/tmp.6KVOpqkFIM + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.7kY0ouRJGU ++ mktemp + local LAST_ERR=/tmp/tmp.sUnsD4Jfqi + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.7kY0ouRJGU + cat /tmp/tmp.sUnsD4Jfqi Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.7kY0ouRJGU /tmp/tmp.sUnsD4Jfqi + return 1 + true + rm -rf /tmp/tmp.WATPD0tkZe