++ echo 'Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/logs/version-service.log' Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/logs/version-service.log ++ '[' -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP= ++ oc get projects error: the server doesn't have a resource type "projects" +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep '\-eks\-' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ '[' ']' +++ kubectl version -o json +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ /usr/bin/sed -r 's/[^0-9.]+//g' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ KUBE_VERSION=1.26 +++ helm version -c +++ /usr/bin/sed -re 's/.*SemVer:"([^"]+)".*/\1/; s/.*\bVersion:"([^"]+)".*/\1/' ++ HELM_VERSION=v3.15.1 ++ '[' v3 == v2 ']' + VS_URL=http://version-service + VS_PORT=11000 + VS_ENDPOINT=http://version-service:11000 + main + create_namespace version-service-7914 + local namespace=version-service-7914 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + cleanup_rbac + awk '{print$1}' + xargs kubectl delete ns + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.a2SGP6q9m2 ++ mktemp + local LAST_OUT=/tmp/tmp.P83OS4yCK1 ++ mktemp + local LAST_ERR=/tmp/tmp.JfGAf4RwDw + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + local LAST_ERR=/tmp/tmp.O5YRNqvn64 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.P83OS4yCK1 + cat /tmp/tmp.O5YRNqvn64 + rm /tmp/tmp.P83OS4yCK1 /tmp/tmp.O5YRNqvn64 + return 0 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level namespace "version-service-13080" deleted + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.a2SGP6q9m2 + cat /tmp/tmp.JfGAf4RwDw Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.a2SGP6q9m2 /tmp/tmp.JfGAf4RwDw + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp + local LAST_OUT=/tmp/tmp.fIygMX5Cuf ++ mktemp + local LAST_ERR=/tmp/tmp.IRo4omNT23 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.fIygMX5Cuf + cat /tmp/tmp.IRo4omNT23 Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.fIygMX5Cuf /tmp/tmp.IRo4omNT23 + return 1 + true + kubectl delete MutatingWebhookConfiguration/chaos-mesh-mutation ValidatingWebhookConfiguration/chaos-mesh-validation ValidatingWebhookConfiguration/validate-auth Error from server (NotFound): mutatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-mutation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-validation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "validate-auth" not found + true + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin delete namespace version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.nKwXffecWw ++ mktemp + local LAST_ERR=/tmp/tmp.yd19FVEYNU + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete namespace version-service-7914 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete namespace version-service-7914 + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete namespace version-service-7914 + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.nKwXffecWw + cat /tmp/tmp.yd19FVEYNU Error from server (NotFound): namespaces "version-service-7914" not found + rm /tmp/tmp.nKwXffecWw /tmp/tmp.yd19FVEYNU + return 1 + : + wait_for_delete namespace/version-service-7914 + local res=namespace/version-service-7914 + set +o xtrace namespace/version-service-7914 - Error from server (NotFound): namespaces "version-service-7914" not found + kubectl_bin create namespace version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.aQOllyuV3B ++ mktemp + local LAST_ERR=/tmp/tmp.Nk13tNw5rI + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl create namespace version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.aQOllyuV3B namespace/version-service-7914 created + cat /tmp/tmp.Nk13tNw5rI + rm /tmp/tmp.aQOllyuV3B /tmp/tmp.Nk13tNw5rI + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.yarqHb86PN +++ mktemp ++ local LAST_ERR=/tmp/tmp.DR7o9FEDIs ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.yarqHb86PN ++ cat /tmp/tmp.DR7o9FEDIs ++ rm /tmp/tmp.yarqHb86PN /tmp/tmp.DR7o9FEDIs ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-version-service --namespace=version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.RYawmLu5CF ++ mktemp + local LAST_ERR=/tmp/tmp.Dyk9hIaOIS + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-version-service --namespace=version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.RYawmLu5CF Context "gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-version-service" modified. + cat /tmp/tmp.Dyk9hIaOIS + rm /tmp/tmp.RYawmLu5CF /tmp/tmp.Dyk9hIaOIS + return 0 + deploy_operator '' /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml disabled install version-service-7914 false http://version-service:11000 + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml + local namespace_mode=disabled + local action=install + local namespace=version-service-7914 + local dont_send_telemetry=false + local version_service_uri_override=http://version-service:11000 + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + prepare_operator_yaml '' /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml disabled install version-service-7914 false + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml + local namespace_mode=disabled + local operator_action=install + local namespace=version-service-7914 + local dont_send_telemetry=false + yq r -d2 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml 'data[values.yaml]' + /usr/bin/sed -e 's#^namespace: .*#namespace: "version-service-7914"#g' + /usr/bin/sed -e 's#^disable_telemetry: .*#disable_telemetry: "false"#g' + /usr/bin/sed -e 's#namespace_mode: .*#namespace_mode: "disabled"#g' + cat /dev/fd/63 - + /usr/bin/sed -e 's#pgo_operator_namespace: .*#pgo_operator_namespace: "version-service-7914"#g' ++ echo 'crunchy_debug: "true"' ++ echo perconalab/percona-postgresql-operator:PR-776-0087ef129 ++ cut -d: -f2 + /usr/bin/sed -e 's#pgo_image_tag: .*#pgo_image_tag: "PR-776-0087ef129"#g' ++ echo perconalab/percona-postgresql-operator:PR-776-0087ef129 ++ cut -d: -f1 + /usr/bin/sed -e 's#pgo_image_prefix: .*#pgo_image_prefix: "perconalab/percona-postgresql-operator"#g' + [[ -n '' ]] + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + yq w -d3 - 'subjects[0].namespace' version-service-7914 + yq w -d4 - 'spec.template.spec.containers[0].image' perconalab/percona-postgresql-operator:PR-776-0087ef129-pgo-deployer + yq w '-d*' /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml metadata.namespace version-service-7914 ++ cat /tmp/tmp.IoigtEzcpK/operator.ini + yq w -d2 - -d2 'data[values.yaml]' 'crunchy_debug: "true" archive_mode: "true" archive_timeout: "60" ccp_image_pull_secret: "" ccp_image_pull_secret_manifest: "" create_rbac: "true" delete_operator_namespace: "false" delete_watched_namespaces: "false" disable_telemetry: "false" namespace: "version-service-7914" namespace_mode: "disabled" pgo_image_prefix: "perconalab/percona-postgresql-operator" pgo_image_pull_policy: "Always" pgo_image_pull_secret: "" pgo_image_pull_secret_manifest: "" pgo_image_tag: "PR-776-0087ef129" pgo_operator_namespace: "version-service-7914"' + yq w -d4 - 'spec.template.spec.containers[0].env[0].value' install + [[ -n '' ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.9HsVPKOPqY ++ mktemp + local LAST_ERR=/tmp/tmp.cNEF8D3xAC + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9HsVPKOPqY serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb created job.batch/pgo-deploy created + cat /tmp/tmp.cNEF8D3xAC + rm /tmp/tmp.9HsVPKOPqY /tmp/tmp.cNEF8D3xAC + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 15 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 16 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 17 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 18 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 19 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 20 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 21 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 22 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 23 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 24 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 25 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 26 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 27 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.IoigtEzcpK/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.77O1vitjMo ++ mktemp + local LAST_ERR=/tmp/tmp.3ZvtJ8uYQN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete -f /tmp/tmp.IoigtEzcpK/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.77O1vitjMo serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.3ZvtJ8uYQN + rm /tmp/tmp.77O1vitjMo /tmp/tmp.3ZvtJ8uYQN + return 0 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TNsy580QK7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.JuThBYoLBT ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.TNsy580QK7 ++ cat /tmp/tmp.JuThBYoLBT ++ rm /tmp/tmp.TNsy580QK7 /tmp/tmp.JuThBYoLBT ++ return 0 + wait_pod postgres-operator-757544477c-g8kbz + local pod=postgres-operator-757544477c-g8kbz + local ns= + set +o xtrace postgres-operator-757544477c-g8kbztrue + [[ -n http://version-service:11000 ]] + kubectl_bin -n version-service-7914 patch deployment/postgres-operator --type json '-p=[{"op":"add","path":"/spec/template/spec/containers/1/env/-","value":{"name":"PERCONA_VS_FALLBACK_URI","value":"http://version-service:11000"}}]' ++ mktemp + local LAST_OUT=/tmp/tmp.yQOWxqL7Ma ++ mktemp + local LAST_ERR=/tmp/tmp.rb7AeSwBWM + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 patch deployment/postgres-operator --type json '-p=[{"op":"add","path":"/spec/template/spec/containers/1/env/-","value":{"name":"PERCONA_VS_FALLBACK_URI","value":"http://version-service:11000"}}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yQOWxqL7Ma deployment.apps/postgres-operator patched + cat /tmp/tmp.rb7AeSwBWM + rm /tmp/tmp.yQOWxqL7Ma /tmp/tmp.rb7AeSwBWM + return 0 + sleep 15 + deploy_version_service + desc 'install version service' + set +o xtrace ----------------------------------------------------------------------------------- install version service ----------------------------------------------------------------------------------- + kubectl_bin create configmap versions --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.dep.json --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.json ++ mktemp + local LAST_OUT=/tmp/tmp.rwu5xAiS1h ++ mktemp + local LAST_ERR=/tmp/tmp.Uh2axDZ2HY + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl create configmap versions --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.dep.json --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.json + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.rwu5xAiS1h configmap/versions created + cat /tmp/tmp.Uh2axDZ2HY + rm /tmp/tmp.rwu5xAiS1h /tmp/tmp.Uh2axDZ2HY + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/vs.yml ++ mktemp + local LAST_OUT=/tmp/tmp.mBcZcTTARx ++ mktemp + local LAST_ERR=/tmp/tmp.fAA7BX1MV2 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/vs.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.mBcZcTTARx deployment.apps/version-service created service/version-service created + cat /tmp/tmp.fAA7BX1MV2 + rm /tmp/tmp.mBcZcTTARx /tmp/tmp.fAA7BX1MV2 + return 0 + /usr/bin/sed s/version-service/version-service-cr/g /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/vs.yml + kubectl_bin apply -f - ++ yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/vs.yml 'spec.template.spec.containers[0].image' ++ mktemp + local LAST_OUT=/tmp/tmp.paOxpjAIkf ++ mktemp + local LAST_ERR=/tmp/tmp.lBmXIpRPD8 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + yq w - -d0 'spec.template.spec.containers[0].image' perconalab/version-service:main-ea25e30 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.paOxpjAIkf deployment.apps/version-service-cr created service/version-service-cr created + cat /tmp/tmp.lBmXIpRPD8 + rm /tmp/tmp.paOxpjAIkf /tmp/tmp.lBmXIpRPD8 + return 0 + sleep 10 + check_telemetry_transfer http://version-service-cr:11000 disabled enabled + local cr_vs_uri=http://version-service-cr:11000 + local cr_vs_channel=disabled + local telemetry_state=enabled + local cluster=minimal + yq r /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr-minimal.yaml + yq w - spec.upgradeOptions.versionServiceEndpoint http://version-service-cr:11000 + yq w - spec.upgradeOptions.apply disabled + spinup_pgcluster minimal /tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml + local cluster=minimal + local config=/tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets minimal version-service-7914 + local cluster=minimal + local namespace=version-service-7914 + yq w - metadata.name minimal-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.jV3HAiCzwj ++ mktemp + local LAST_ERR=/tmp/tmp.mxgyNlXFFK + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.jV3HAiCzwj secret/minimal-users-secret created + cat /tmp/tmp.mxgyNlXFFK + rm /tmp/tmp.jV3HAiCzwj /tmp/tmp.mxgyNlXFFK + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name minimal-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.j9KiPpGZAC ++ mktemp + local LAST_ERR=/tmp/tmp.r85GHqzqXG + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.j9KiPpGZAC secret/minimal-pmm-secret created + cat /tmp/tmp.r85GHqzqXG + rm /tmp/tmp.j9KiPpGZAC /tmp/tmp.r85GHqzqXG + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.mhRV7PRA6t ++ mktemp + local LAST_ERR=/tmp/tmp.xe96PzltZb + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.mhRV7PRA6t deployment.apps/pg-client created + cat /tmp/tmp.xe96PzltZb + rm /tmp/tmp.mhRV7PRA6t /tmp/tmp.xe96PzltZb + return 0 + apply_cluster /tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml minimal false false false false false false false false + local path=/tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml + local name=minimal + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml metadata.annotations.current-primary minimal + yq w - metadata.labels.deployment-name minimal + yq w - metadata.labels.name minimal + yq w - spec.database minimal + yq w - spec.user some-name + yq w - spec.name minimal + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.keepData false + yq w - metadata.labels.pg-cluster minimal + yq w - spec.keepBackups false + yq w - spec.secretsName minimal-users-secret + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.clustername minimal + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - metadata.name minimal + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - metadata.labels.crunchy-pgha-scope minimal + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.rQXGvh4I9Y ++ mktemp + local LAST_ERR=/tmp/tmp.giqxT53Vwh + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.rQXGvh4I9Y perconapgcluster.pg.percona.com/minimal created + cat /tmp/tmp.giqxT53Vwh + rm /tmp/tmp.rQXGvh4I9Y /tmp/tmp.giqxT53Vwh + return 0 + wait_deployment minimal-backrest-shared-repo + local name=minimal-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace minimal-backrest-shared-repo. + wait_deployment minimal + local name=minimal + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace minimal........ + [[ false == \g\c\s\+* ]] + wait_job_completion minimal-stanza-create false + local job=minimal-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' minimal-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-minimal + local job=backrest-backup-minimal + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' backrest-backup-minimal+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency minimal + cluster_name=minimal + wait_cluster_status minimal 'pgcluster Initialized' + local cluster_name=minimal + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eIpUu4ncIM +++ mktemp ++ local LAST_ERR=/tmp/tmp.7hPxpwMeVu ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.eIpUu4ncIM ++ cat /tmp/tmp.7hPxpwMeVu ++ rm /tmp/tmp.eIpUu4ncIM /tmp/tmp.7hPxpwMeVu ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Pgg07gemwF ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ojmKyOBHgU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Pgg07gemwF +++ cat /tmp/tmp.ojmKyOBHgU +++ rm /tmp/tmp.Pgg07gemwF /tmp/tmp.ojmKyOBHgU +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mj5O9cmOOm ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jlAXYBdfQF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.mj5O9cmOOm +++ cat /tmp/tmp.jlAXYBdfQF +++ rm /tmp/tmp.mj5O9cmOOm /tmp/tmp.jlAXYBdfQF +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dkP8kpBR11 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.n5UbNA1ctw +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.dkP8kpBR11 +++ cat /tmp/tmp.n5UbNA1ctw +++ rm /tmp/tmp.dkP8kpBR11 /tmp/tmp.n5UbNA1ctw +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.uPTHO2UtaP ++++ mktemp +++ local LAST_ERR=/tmp/tmp.aUP4haAtjN +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.uPTHO2UtaP +++ cat /tmp/tmp.aUP4haAtjN +++ rm /tmp/tmp.uPTHO2UtaP /tmp/tmp.aUP4haAtjN +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.M43T8sQC91 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.dAEuMOm7mE +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.M43T8sQC91 +++ cat /tmp/tmp.dAEuMOm7mE +++ rm /tmp/tmp.M43T8sQC91 /tmp/tmp.dAEuMOm7mE +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.XCgCTRHZUv +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.3UTNGuBA5T ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.XCgCTRHZUv ++++ cat /tmp/tmp.3UTNGuBA5T ++++ rm /tmp/tmp.XCgCTRHZUv /tmp/tmp.3UTNGuBA5T ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.589XMPHYux ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ig9Jcc39rQ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.589XMPHYux +++ cat /tmp/tmp.ig9Jcc39rQ +++ rm /tmp/tmp.589XMPHYux /tmp/tmp.ig9Jcc39rQ +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@minimal.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VyViWjFYHQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.jHDwBqi7b3 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.VyViWjFYHQ ++ cat /tmp/tmp.jHDwBqi7b3 ++ rm /tmp/tmp.VyViWjFYHQ /tmp/tmp.jHDwBqi7b3 ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.hf7ghhq7V8 ++ mktemp + local LAST_ERR=/tmp/tmp.VsEh814XVG + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.hf7ghhq7V8 + cat /tmp/tmp.VsEh814XVG + rm /tmp/tmp.hf7ghhq7V8 /tmp/tmp.VsEh814XVG + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ifuyVq22jn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.B2yrzZ2OT1 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ifuyVq22jn +++ cat /tmp/tmp.B2yrzZ2OT1 +++ rm /tmp/tmp.ifuyVq22jn /tmp/tmp.B2yrzZ2OT1 +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.EkOjskRD2M ++++ mktemp +++ local LAST_ERR=/tmp/tmp.IvIDW5oGbz +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.EkOjskRD2M +++ cat /tmp/tmp.IvIDW5oGbz +++ rm /tmp/tmp.EkOjskRD2M /tmp/tmp.IvIDW5oGbz +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.jHgsFStbL0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.GizytrEi6Z +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.jHgsFStbL0 +++ cat /tmp/tmp.GizytrEi6Z +++ rm /tmp/tmp.jHgsFStbL0 /tmp/tmp.GizytrEi6Z +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.brMr8YePBy +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.DdtEmfJW47 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.brMr8YePBy ++++ cat /tmp/tmp.DdtEmfJW47 ++++ rm /tmp/tmp.brMr8YePBy /tmp/tmp.DdtEmfJW47 ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.GlHUS5onfe ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ie9DCcICAK +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.GlHUS5onfe +++ cat /tmp/tmp.ie9DCcICAK +++ rm /tmp/tmp.GlHUS5onfe /tmp/tmp.ie9DCcICAK +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@minimal.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I2p3WJ9HTG +++ mktemp ++ local LAST_ERR=/tmp/tmp.JNseyZ5fz5 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.I2p3WJ9HTG ++ cat /tmp/tmp.JNseyZ5fz5 ++ rm /tmp/tmp.I2p3WJ9HTG /tmp/tmp.JNseyZ5fz5 ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.FnNeszlVIg ++ mktemp + local LAST_ERR=/tmp/tmp.CP92iHL6nt + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FnNeszlVIg + cat /tmp/tmp.CP92iHL6nt + rm /tmp/tmp.FnNeszlVIg /tmp/tmp.CP92iHL6nt + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hmOQQWyy85 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.EKfBf66Yz0 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.hmOQQWyy85 +++ cat /tmp/tmp.EKfBf66Yz0 +++ rm /tmp/tmp.hmOQQWyy85 /tmp/tmp.EKfBf66Yz0 +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OTPhn8NT0k ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jkJPQhYzxX +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.OTPhn8NT0k +++ cat /tmp/tmp.jkJPQhYzxX +++ rm /tmp/tmp.OTPhn8NT0k /tmp/tmp.jkJPQhYzxX +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ASRoRKqu2a ++++ mktemp +++ local LAST_ERR=/tmp/tmp.AVLcZjuiWp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ASRoRKqu2a +++ cat /tmp/tmp.AVLcZjuiWp +++ rm /tmp/tmp.ASRoRKqu2a /tmp/tmp.AVLcZjuiWp +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.QoymZhR2G5 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.VzkETVA6Ls ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.QoymZhR2G5 ++++ cat /tmp/tmp.VzkETVA6Ls ++++ rm /tmp/tmp.QoymZhR2G5 /tmp/tmp.VzkETVA6Ls ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ASZU5oq689 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.IBzGhx3ezG +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ASZU5oq689 +++ cat /tmp/tmp.IBzGhx3ezG +++ rm /tmp/tmp.ASZU5oq689 /tmp/tmp.IBzGhx3ezG +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@minimal.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GsNbNWvmF7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HwuAJDZgxd ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.GsNbNWvmF7 ++ cat /tmp/tmp.HwuAJDZgxd ++ rm /tmp/tmp.GsNbNWvmF7 /tmp/tmp.HwuAJDZgxd ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.ojzVzfxBUr ++ mktemp + local LAST_ERR=/tmp/tmp.DnLuCwyLdh + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ojzVzfxBUr + cat /tmp/tmp.DnLuCwyLdh + rm /tmp/tmp.ojzVzfxBUr /tmp/tmp.DnLuCwyLdh + return 0 + sleep 10 + grep -Eo '\{.*\}' ++ kubectl get pods --selector=run=version-service-cr -o 'jsonpath={.items[0].metadata.name}' + jq 'del(."grpc.request.content".msg.customResourceUid)' + grep -E 'server request payload|unary call' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.start_time")' + kubectl_bin logs version-service-cr-76cc4777bf-w7w5n ++ mktemp + local LAST_OUT=/tmp/tmp.BU2pvyi2Gk ++ mktemp + local LAST_ERR=/tmp/tmp.XcUV8xlg7p + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl logs version-service-cr-76cc4777bf-w7w5n + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.BU2pvyi2Gk + cat /tmp/tmp.XcUV8xlg7p + rm /tmp/tmp.BU2pvyi2Gk /tmp/tmp.XcUV8xlg7p + return 0 + grep -E 'server request payload|unary call' + grep -Eo '\{.*\}' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.request.content".msg.customResourceUid)' + jq 'del(."grpc.start_time")' ++ kubectl get pods --selector=run=version-service -o 'jsonpath={.items[0].metadata.name}' + kubectl_bin logs version-service-5cd4db57-ss744 ++ mktemp + local LAST_OUT=/tmp/tmp.QQe2Y4WEtL ++ mktemp + local LAST_ERR=/tmp/tmp.5xya2DEmCB + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl logs version-service-5cd4db57-ss744 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.QQe2Y4WEtL + cat /tmp/tmp.5xya2DEmCB + rm /tmp/tmp.QQe2Y4WEtL /tmp/tmp.5xya2DEmCB + return 0 + diff -u /tmp/tmp.IoigtEzcpK/enabled_telemetry.version-service.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/enabled_telemetry.version-service.log.json + [[ enabled == \d\i\s\a\b\l\e\d ]] + diff -u /tmp/tmp.IoigtEzcpK/enabled_telemetry.version-service-cr.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/enabled_telemetry.version-service-cr.log.json + kubectl_bin delete perconapgcluster/minimal ++ mktemp + local LAST_OUT=/tmp/tmp.8QmSgkttEz ++ mktemp + local LAST_ERR=/tmp/tmp.ss304dmct9 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete perconapgcluster/minimal + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.8QmSgkttEz perconapgcluster.pg.percona.com "minimal" deleted + cat /tmp/tmp.ss304dmct9 + rm /tmp/tmp.8QmSgkttEz /tmp/tmp.ss304dmct9 + return 0 ++ kubectl_bin -n version-service-7914 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ISBEQsqPmv +++ mktemp ++ local LAST_ERR=/tmp/tmp.pJVVwb5CR0 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.ISBEQsqPmv ++ cat /tmp/tmp.pJVVwb5CR0 ++ rm /tmp/tmp.ISBEQsqPmv /tmp/tmp.pJVVwb5CR0 ++ return 0 + wait_job_completion minimal-rmdata-mava + local job=minimal-rmdata-mava + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' minimal-rmdata-mava+ retry=0 ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-mava -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + kubectl_bin -n version-service-7914 delete pod --selector=run=version-service ++ mktemp + local LAST_OUT=/tmp/tmp.Ab0pdcvdIX ++ mktemp + local LAST_ERR=/tmp/tmp.m5gCG8XHoH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 delete pod --selector=run=version-service + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Ab0pdcvdIX pod "version-service-5cd4db57-ss744" deleted + cat /tmp/tmp.m5gCG8XHoH + rm /tmp/tmp.Ab0pdcvdIX /tmp/tmp.m5gCG8XHoH + return 0 + wait_deployment version-service + local name=version-service + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace version-service + kubectl_bin -n version-service-7914 delete pod --selector=run=version-service-cr ++ mktemp + local LAST_OUT=/tmp/tmp.yFWgwSOroR ++ mktemp + local LAST_ERR=/tmp/tmp.VedTmuGbUG + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 delete pod --selector=run=version-service-cr + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yFWgwSOroR pod "version-service-cr-76cc4777bf-w7w5n" deleted + cat /tmp/tmp.VedTmuGbUG + rm /tmp/tmp.yFWgwSOroR /tmp/tmp.VedTmuGbUG + return 0 + wait_deployment version-service-cr + local name=version-service-cr + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace version-service-cr + kubectl_bin -n version-service-7914 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"true"}}]' ++ mktemp + local LAST_OUT=/tmp/tmp.85kHBdrju4 ++ mktemp + local LAST_ERR=/tmp/tmp.k7MkVkhdST + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"true"}}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.85kHBdrju4 deployment.apps/postgres-operator patched + cat /tmp/tmp.k7MkVkhdST + rm /tmp/tmp.85kHBdrju4 /tmp/tmp.k7MkVkhdST + return 0 + wait_deployment postgres-operator + local name=postgres-operator + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace postgres-operator.... + check_telemetry_transfer http://version-service-cr:11000 disabled disabled + local cr_vs_uri=http://version-service-cr:11000 + local cr_vs_channel=disabled + local telemetry_state=disabled + local cluster=minimal + yq r /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr-minimal.yaml + yq w - spec.upgradeOptions.apply disabled + yq w - spec.upgradeOptions.versionServiceEndpoint http://version-service-cr:11000 + spinup_pgcluster minimal /tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml + local cluster=minimal + local config=/tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets minimal version-service-7914 + local cluster=minimal + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name minimal-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.hkQcVvecpj ++ mktemp + local LAST_ERR=/tmp/tmp.S9OJxuCjBl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.hkQcVvecpj secret/minimal-users-secret created + cat /tmp/tmp.S9OJxuCjBl + rm /tmp/tmp.hkQcVvecpj /tmp/tmp.S9OJxuCjBl + return 0 + yq w - metadata.name minimal-pmm-secret + kubectl_bin -n version-service-7914 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.QdItYVA7KG ++ mktemp + local LAST_ERR=/tmp/tmp.Ylep3bobeQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.QdItYVA7KG secret/minimal-pmm-secret configured + cat /tmp/tmp.Ylep3bobeQ + rm /tmp/tmp.QdItYVA7KG /tmp/tmp.Ylep3bobeQ + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.M8wyodraBb ++ mktemp + local LAST_ERR=/tmp/tmp.Tz640y24qb + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.M8wyodraBb deployment.apps/pg-client unchanged + cat /tmp/tmp.Tz640y24qb + rm /tmp/tmp.M8wyodraBb /tmp/tmp.Tz640y24qb + return 0 + apply_cluster /tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml minimal false false false false false false false false + local path=/tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml + local name=minimal + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /tmp/tmp.IoigtEzcpK/cr.minimal.apply.disabled.yaml metadata.annotations.current-primary minimal + yq w - metadata.labels.crunchy-pgha-scope minimal + yq w - metadata.labels.pg-cluster minimal + yq w - metadata.name minimal + yq w - metadata.labels.name minimal + yq w - spec.user some-name + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.clustername minimal + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.database minimal + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.name minimal + yq w - spec.keepBackups false + yq w - spec.secretsName minimal-users-secret + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.keepData false + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - metadata.labels.deployment-name minimal + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.X1hb4QSuMe ++ mktemp + local LAST_ERR=/tmp/tmp.etu7ST18eg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.X1hb4QSuMe perconapgcluster.pg.percona.com/minimal created + cat /tmp/tmp.etu7ST18eg + rm /tmp/tmp.X1hb4QSuMe /tmp/tmp.etu7ST18eg + return 0 + wait_deployment minimal-backrest-shared-repo + local name=minimal-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace minimal-backrest-shared-repo... + wait_deployment minimal + local name=minimal + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace minimal....... + [[ false == \g\c\s\+* ]] + wait_job_completion minimal-stanza-create false + local job=minimal-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' minimal-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-minimal + local job=backrest-backup-minimal + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' backrest-backup-minimal+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency minimal + cluster_name=minimal + wait_cluster_status minimal 'pgcluster Initialized' + local cluster_name=minimal + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gYnDz59TRS +++ mktemp ++ local LAST_ERR=/tmp/tmp.DgprIZEWtU ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.gYnDz59TRS ++ cat /tmp/tmp.DgprIZEWtU ++ rm /tmp/tmp.gYnDz59TRS /tmp/tmp.DgprIZEWtU ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pmRuTFU2Fr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.X512y3ELQ6 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.pmRuTFU2Fr +++ cat /tmp/tmp.X512y3ELQ6 +++ rm /tmp/tmp.pmRuTFU2Fr /tmp/tmp.X512y3ELQ6 +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Pr5p0SEfeg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fuTKQZIJ35 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Pr5p0SEfeg +++ cat /tmp/tmp.fuTKQZIJ35 +++ rm /tmp/tmp.Pr5p0SEfeg /tmp/tmp.fuTKQZIJ35 +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qNkUOsp5Q0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.xV57oaHOeG +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qNkUOsp5Q0 +++ cat /tmp/tmp.xV57oaHOeG +++ rm /tmp/tmp.qNkUOsp5Q0 /tmp/tmp.xV57oaHOeG +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0D1dXxJA5K ++++ mktemp +++ local LAST_ERR=/tmp/tmp.n9owC2KaC6 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.0D1dXxJA5K +++ cat /tmp/tmp.n9owC2KaC6 +++ rm /tmp/tmp.0D1dXxJA5K /tmp/tmp.n9owC2KaC6 +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.i45RYDEFSc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.hsNp0Sut8q +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.i45RYDEFSc +++ cat /tmp/tmp.hsNp0Sut8q +++ rm /tmp/tmp.i45RYDEFSc /tmp/tmp.hsNp0Sut8q +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.KeJf7Lr0hU +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.55HIuBQi4A ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.KeJf7Lr0hU ++++ cat /tmp/tmp.55HIuBQi4A ++++ rm /tmp/tmp.KeJf7Lr0hU /tmp/tmp.55HIuBQi4A ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.T9QwAVr9DH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XxdalvePlH +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.T9QwAVr9DH +++ cat /tmp/tmp.XxdalvePlH +++ rm /tmp/tmp.T9QwAVr9DH /tmp/tmp.XxdalvePlH +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@minimal.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e92oke27FV +++ mktemp ++ local LAST_ERR=/tmp/tmp.TNK4Lg20Ao ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.e92oke27FV ++ cat /tmp/tmp.TNK4Lg20Ao ++ rm /tmp/tmp.e92oke27FV /tmp/tmp.TNK4Lg20Ao ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.NwmXS9jpI2 ++ mktemp + local LAST_ERR=/tmp/tmp.2lJy7FLUml + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.NwmXS9jpI2 + cat /tmp/tmp.2lJy7FLUml + rm /tmp/tmp.NwmXS9jpI2 /tmp/tmp.2lJy7FLUml + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.h27VwSVxJQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gk02crTPCu +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.h27VwSVxJQ +++ cat /tmp/tmp.gk02crTPCu +++ rm /tmp/tmp.h27VwSVxJQ /tmp/tmp.gk02crTPCu +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.JkBaoXufmn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gJersUL7nO +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.JkBaoXufmn +++ cat /tmp/tmp.gJersUL7nO +++ rm /tmp/tmp.JkBaoXufmn /tmp/tmp.gJersUL7nO +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MoB9HBfKrq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fBchBXWfzQ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MoB9HBfKrq +++ cat /tmp/tmp.fBchBXWfzQ +++ rm /tmp/tmp.MoB9HBfKrq /tmp/tmp.fBchBXWfzQ +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.CtPVMWpXPZ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.7ZLmRZd0Oi ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.CtPVMWpXPZ ++++ cat /tmp/tmp.7ZLmRZd0Oi ++++ rm /tmp/tmp.CtPVMWpXPZ /tmp/tmp.7ZLmRZd0Oi ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZZIrdN13YO ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LysPwqki60 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZZIrdN13YO +++ cat /tmp/tmp.LysPwqki60 +++ rm /tmp/tmp.ZZIrdN13YO /tmp/tmp.LysPwqki60 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@minimal.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kNPgBm0fDk +++ mktemp ++ local LAST_ERR=/tmp/tmp.KAGGc3Gkb2 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.kNPgBm0fDk ++ cat /tmp/tmp.KAGGc3Gkb2 ++ rm /tmp/tmp.kNPgBm0fDk /tmp/tmp.KAGGc3Gkb2 ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.xydSlqYCeH ++ mktemp + local LAST_ERR=/tmp/tmp.RxHp2lAi4y + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xydSlqYCeH + cat /tmp/tmp.RxHp2lAi4y + rm /tmp/tmp.xydSlqYCeH /tmp/tmp.RxHp2lAi4y + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ET5VBnmOGF ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZxBsp3snVr +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ET5VBnmOGF +++ cat /tmp/tmp.ZxBsp3snVr +++ rm /tmp/tmp.ET5VBnmOGF /tmp/tmp.ZxBsp3snVr +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Wt7e1752c0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.tXYbIaDV0C +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Wt7e1752c0 +++ cat /tmp/tmp.tXYbIaDV0C +++ rm /tmp/tmp.Wt7e1752c0 /tmp/tmp.tXYbIaDV0C +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.onDpRMn193 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.eCRFW4HuJP +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.onDpRMn193 +++ cat /tmp/tmp.eCRFW4HuJP +++ rm /tmp/tmp.onDpRMn193 /tmp/tmp.eCRFW4HuJP +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.TXke4aKUJX +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.m1U02aswVW ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.TXke4aKUJX ++++ cat /tmp/tmp.m1U02aswVW ++++ rm /tmp/tmp.TXke4aKUJX /tmp/tmp.m1U02aswVW ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.kGpS58esy2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.F6P6ckRqG6 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.kGpS58esy2 +++ cat /tmp/tmp.F6P6ckRqG6 +++ rm /tmp/tmp.kGpS58esy2 /tmp/tmp.F6P6ckRqG6 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@minimal.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dO0uE1FXaW +++ mktemp ++ local LAST_ERR=/tmp/tmp.g76OVjtd8L ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.dO0uE1FXaW ++ cat /tmp/tmp.g76OVjtd8L ++ rm /tmp/tmp.dO0uE1FXaW /tmp/tmp.g76OVjtd8L ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.OhfJkXUZbs ++ mktemp + local LAST_ERR=/tmp/tmp.W0mdI52KKK + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.OhfJkXUZbs + cat /tmp/tmp.W0mdI52KKK + rm /tmp/tmp.OhfJkXUZbs /tmp/tmp.W0mdI52KKK + return 0 + sleep 10 + grep -E 'server request payload|unary call' + grep -Eo '\{.*\}' ++ kubectl get pods --selector=run=version-service-cr -o 'jsonpath={.items[0].metadata.name}' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.start_time")' + jq 'del(."grpc.request.content".msg.customResourceUid)' + kubectl_bin logs version-service-cr-76cc4777bf-vgms9 ++ mktemp + local LAST_OUT=/tmp/tmp.PfspdvKN8a ++ mktemp + local LAST_ERR=/tmp/tmp.aeHxgEWKjl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl logs version-service-cr-76cc4777bf-vgms9 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PfspdvKN8a + cat /tmp/tmp.aeHxgEWKjl + rm /tmp/tmp.PfspdvKN8a /tmp/tmp.aeHxgEWKjl + return 0 + grep -E 'server request payload|unary call' + grep -Eo '\{.*\}' ++ kubectl get pods --selector=run=version-service -o 'jsonpath={.items[0].metadata.name}' + jq 'del(."grpc.request.content".msg.customResourceUid)' + jq 'del(."grpc.start_time")' + jq 'del(."grpc.time_ms")' + kubectl_bin logs version-service-5cd4db57-vtrrk ++ mktemp + local LAST_OUT=/tmp/tmp.6Hrz8GWR9g ++ mktemp + local LAST_ERR=/tmp/tmp.s2QUOUdjZE + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl logs version-service-5cd4db57-vtrrk + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.6Hrz8GWR9g + cat /tmp/tmp.s2QUOUdjZE + rm /tmp/tmp.6Hrz8GWR9g /tmp/tmp.s2QUOUdjZE + return 0 + diff -u /tmp/tmp.IoigtEzcpK/disabled_telemetry.version-service.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/disabled_telemetry.version-service.log.json + [[ disabled == \d\i\s\a\b\l\e\d ]] + [[ disabled == \r\e\c\o\m\m\e\n\d\e\d ]] + diff -u /tmp/tmp.IoigtEzcpK/disabled_telemetry.version-service-cr.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/disabled_telemetry.version-service-cr.log.json + kubectl_bin delete perconapgcluster/minimal ++ mktemp + local LAST_OUT=/tmp/tmp.ZgQMoLuPoO ++ mktemp + local LAST_ERR=/tmp/tmp.RyTKEwzDtf + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete perconapgcluster/minimal + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ZgQMoLuPoO perconapgcluster.pg.percona.com "minimal" deleted + cat /tmp/tmp.RyTKEwzDtf + rm /tmp/tmp.ZgQMoLuPoO /tmp/tmp.RyTKEwzDtf + return 0 ++ kubectl_bin -n version-service-7914 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MT8KobqsmT +++ mktemp ++ local LAST_ERR=/tmp/tmp.UPTQYAOcHf ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.MT8KobqsmT ++ cat /tmp/tmp.UPTQYAOcHf ++ rm /tmp/tmp.MT8KobqsmT /tmp/tmp.UPTQYAOcHf ++ return 0 + wait_job_completion minimal-rmdata-wqka + local job=minimal-rmdata-wqka + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' minimal-rmdata-wqka+ retry=0 ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-wqka -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + check_telemetry_transfer http://version-service-cr:11000 recommended disabled + local cr_vs_uri=http://version-service-cr:11000 + local cr_vs_channel=recommended + local telemetry_state=disabled + local cluster=minimal + yq r /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr-minimal.yaml + yq w - spec.upgradeOptions.apply recommended + yq w - spec.upgradeOptions.versionServiceEndpoint http://version-service-cr:11000 + spinup_pgcluster minimal /tmp/tmp.IoigtEzcpK/cr.minimal.apply.recommended.yaml + local cluster=minimal + local config=/tmp/tmp.IoigtEzcpK/cr.minimal.apply.recommended.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets minimal version-service-7914 + local cluster=minimal + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name minimal-users-secret + kubectl_bin -n version-service-7914 apply -f - + yq w - stringData.some-name some-name_pass ++ mktemp + local LAST_OUT=/tmp/tmp.BixbRb9CMP ++ mktemp + local LAST_ERR=/tmp/tmp.2mRswwkdAm + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.BixbRb9CMP secret/minimal-users-secret created + cat /tmp/tmp.2mRswwkdAm + rm /tmp/tmp.BixbRb9CMP /tmp/tmp.2mRswwkdAm + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name minimal-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.HNdj4kP3ON ++ mktemp + local LAST_ERR=/tmp/tmp.7cZNLpRdsY + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.HNdj4kP3ON secret/minimal-pmm-secret configured + cat /tmp/tmp.7cZNLpRdsY + rm /tmp/tmp.HNdj4kP3ON /tmp/tmp.7cZNLpRdsY + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.QNrgWjrDb9 ++ mktemp + local LAST_ERR=/tmp/tmp.4gApXJmXUl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.QNrgWjrDb9 deployment.apps/pg-client unchanged + cat /tmp/tmp.4gApXJmXUl + rm /tmp/tmp.QNrgWjrDb9 /tmp/tmp.4gApXJmXUl + return 0 + apply_cluster /tmp/tmp.IoigtEzcpK/cr.minimal.apply.recommended.yaml minimal false false false false false false false false + local path=/tmp/tmp.IoigtEzcpK/cr.minimal.apply.recommended.yaml + local name=minimal + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /tmp/tmp.IoigtEzcpK/cr.minimal.apply.recommended.yaml metadata.annotations.current-primary minimal + yq w - metadata.labels.crunchy-pgha-scope minimal + yq w - metadata.labels.deployment-name minimal + yq w - metadata.labels.pg-cluster minimal + yq w - spec.clustername minimal + yq w - spec.database minimal + yq w - metadata.name minimal + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.name minimal + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - metadata.labels.name minimal + yq w - spec.secretsName minimal-users-secret + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.keepData false + yq w - spec.keepBackups false + yq w - spec.user some-name + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.mlg0FcUmrL ++ mktemp + local LAST_ERR=/tmp/tmp.KeQR62UQxD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.mlg0FcUmrL perconapgcluster.pg.percona.com/minimal created + cat /tmp/tmp.KeQR62UQxD + rm /tmp/tmp.mlg0FcUmrL /tmp/tmp.KeQR62UQxD + return 0 + wait_deployment minimal-backrest-shared-repo + local name=minimal-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace minimal-backrest-shared-repo.. + wait_deployment minimal + local name=minimal + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace minimal....... + [[ false == \g\c\s\+* ]] + wait_job_completion minimal-stanza-create false + local job=minimal-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' minimal-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-minimal + local job=backrest-backup-minimal + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' backrest-backup-minimal+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency minimal + cluster_name=minimal + wait_cluster_status minimal 'pgcluster Initialized' + local cluster_name=minimal + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.opjJypClzi +++ mktemp ++ local LAST_ERR=/tmp/tmp.czp54Icd4e ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.opjJypClzi ++ cat /tmp/tmp.czp54Icd4e ++ rm /tmp/tmp.opjJypClzi /tmp/tmp.czp54Icd4e ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1au2UNP9iG ++++ mktemp +++ local LAST_ERR=/tmp/tmp.udLZ5Kph1F +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.1au2UNP9iG +++ cat /tmp/tmp.udLZ5Kph1F +++ rm /tmp/tmp.1au2UNP9iG /tmp/tmp.udLZ5Kph1F +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.PgIYoKJT0d ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zNqjeRwJmg +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.PgIYoKJT0d +++ cat /tmp/tmp.zNqjeRwJmg +++ rm /tmp/tmp.PgIYoKJT0d /tmp/tmp.zNqjeRwJmg +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.eXWQqGRcS2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gAzjrIXtcl +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.eXWQqGRcS2 +++ cat /tmp/tmp.gAzjrIXtcl +++ rm /tmp/tmp.eXWQqGRcS2 /tmp/tmp.gAzjrIXtcl +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mC6s2ADGMj ++++ mktemp +++ local LAST_ERR=/tmp/tmp.OlYfVUPeyH +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.mC6s2ADGMj +++ cat /tmp/tmp.OlYfVUPeyH +++ rm /tmp/tmp.mC6s2ADGMj /tmp/tmp.OlYfVUPeyH +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.R8Tx648O6p ++++ mktemp +++ local LAST_ERR=/tmp/tmp.35UDKvcwux +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.R8Tx648O6p +++ cat /tmp/tmp.35UDKvcwux +++ rm /tmp/tmp.R8Tx648O6p /tmp/tmp.35UDKvcwux +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Ie861oBZCs +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.9R3WleAzbN ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.Ie861oBZCs ++++ cat /tmp/tmp.9R3WleAzbN ++++ rm /tmp/tmp.Ie861oBZCs /tmp/tmp.9R3WleAzbN ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.R7VH7J1TWX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.xIYfao723h +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.R7VH7J1TWX +++ cat /tmp/tmp.xIYfao723h +++ rm /tmp/tmp.R7VH7J1TWX /tmp/tmp.xIYfao723h +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@minimal.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aVfYCuhK75 +++ mktemp ++ local LAST_ERR=/tmp/tmp.7TR7Ak4har ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.aVfYCuhK75 ++ cat /tmp/tmp.7TR7Ak4har ++ rm /tmp/tmp.aVfYCuhK75 /tmp/tmp.7TR7Ak4har ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.PWvgp4m8eC ++ mktemp + local LAST_ERR=/tmp/tmp.ZicFZQtoMp + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PWvgp4m8eC + cat /tmp/tmp.ZicFZQtoMp + rm /tmp/tmp.PWvgp4m8eC /tmp/tmp.ZicFZQtoMp + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZlFYk1Eb4G ++++ mktemp +++ local LAST_ERR=/tmp/tmp.L09IaDRmuG +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZlFYk1Eb4G +++ cat /tmp/tmp.L09IaDRmuG +++ rm /tmp/tmp.ZlFYk1Eb4G /tmp/tmp.L09IaDRmuG +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.abSVtfyzbo ++++ mktemp +++ local LAST_ERR=/tmp/tmp.EyplJtZ0GV +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.abSVtfyzbo +++ cat /tmp/tmp.EyplJtZ0GV +++ rm /tmp/tmp.abSVtfyzbo /tmp/tmp.EyplJtZ0GV +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hS8cdIdSy2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NuB8ePJThB +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.hS8cdIdSy2 +++ cat /tmp/tmp.NuB8ePJThB +++ rm /tmp/tmp.hS8cdIdSy2 /tmp/tmp.NuB8ePJThB +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.pFty8jZjkw +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.2uwfpueFWz ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.pFty8jZjkw ++++ cat /tmp/tmp.2uwfpueFWz ++++ rm /tmp/tmp.pFty8jZjkw /tmp/tmp.2uwfpueFWz ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.329Ke9H1DV ++++ mktemp +++ local LAST_ERR=/tmp/tmp.OEcg28YPGm +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.329Ke9H1DV +++ cat /tmp/tmp.OEcg28YPGm +++ rm /tmp/tmp.329Ke9H1DV /tmp/tmp.OEcg28YPGm +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@minimal.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CDJXKomyHL +++ mktemp ++ local LAST_ERR=/tmp/tmp.fuBSe2JoZg ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.CDJXKomyHL ++ cat /tmp/tmp.fuBSe2JoZg ++ rm /tmp/tmp.CDJXKomyHL /tmp/tmp.fuBSe2JoZg ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.apmSS27fCV ++ mktemp + local LAST_ERR=/tmp/tmp.xld3N7CKRW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.apmSS27fCV + cat /tmp/tmp.xld3N7CKRW + rm /tmp/tmp.apmSS27fCV /tmp/tmp.xld3N7CKRW + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.EtAejWQQfu ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BXxeEBEO5N +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.EtAejWQQfu +++ cat /tmp/tmp.BXxeEBEO5N +++ rm /tmp/tmp.EtAejWQQfu /tmp/tmp.BXxeEBEO5N +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1hHyo3GfEC ++++ mktemp +++ local LAST_ERR=/tmp/tmp.FZuq3KqWIQ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.1hHyo3GfEC +++ cat /tmp/tmp.FZuq3KqWIQ +++ rm /tmp/tmp.1hHyo3GfEC /tmp/tmp.FZuq3KqWIQ +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZV3UvWl6nQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jtqtmGQ5dF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZV3UvWl6nQ +++ cat /tmp/tmp.jtqtmGQ5dF +++ rm /tmp/tmp.ZV3UvWl6nQ /tmp/tmp.jtqtmGQ5dF +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.yaG4dnmaXo +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.87KwHBYgXn ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.yaG4dnmaXo ++++ cat /tmp/tmp.87KwHBYgXn ++++ rm /tmp/tmp.yaG4dnmaXo /tmp/tmp.87KwHBYgXn ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DrZZ9s3l5E ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Z4xpj6KX1A +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.DrZZ9s3l5E +++ cat /tmp/tmp.Z4xpj6KX1A +++ rm /tmp/tmp.DrZZ9s3l5E /tmp/tmp.Z4xpj6KX1A +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@minimal.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@minimal.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PvnClOAzVg +++ mktemp ++ local LAST_ERR=/tmp/tmp.WjYj1BhH20 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.PvnClOAzVg ++ cat /tmp/tmp.WjYj1BhH20 ++ rm /tmp/tmp.PvnClOAzVg /tmp/tmp.WjYj1BhH20 ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.bYEMnlLYOj ++ mktemp + local LAST_ERR=/tmp/tmp.zZO9RyGLTd + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.bYEMnlLYOj + cat /tmp/tmp.zZO9RyGLTd + rm /tmp/tmp.bYEMnlLYOj /tmp/tmp.zZO9RyGLTd + return 0 + sleep 10 + grep -E 'server request payload|unary call' + grep -Eo '\{.*\}' ++ kubectl get pods --selector=run=version-service-cr -o 'jsonpath={.items[0].metadata.name}' + jq 'del(."grpc.request.content".msg.customResourceUid)' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.start_time")' + kubectl_bin logs version-service-cr-76cc4777bf-vgms9 ++ mktemp + local LAST_OUT=/tmp/tmp.cjbsRqHBzY ++ mktemp + local LAST_ERR=/tmp/tmp.yHVtFHSDVQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl logs version-service-cr-76cc4777bf-vgms9 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.cjbsRqHBzY + cat /tmp/tmp.yHVtFHSDVQ + rm /tmp/tmp.cjbsRqHBzY /tmp/tmp.yHVtFHSDVQ + return 0 + grep -E 'server request payload|unary call' + jq 'del(."grpc.request.content".msg.customResourceUid)' + jq 'del(."grpc.start_time")' ++ kubectl get pods --selector=run=version-service -o 'jsonpath={.items[0].metadata.name}' + grep -Eo '\{.*\}' + jq 'del(."grpc.time_ms")' + kubectl_bin logs version-service-5cd4db57-vtrrk ++ mktemp + local LAST_OUT=/tmp/tmp.1ZMccy00iz ++ mktemp + local LAST_ERR=/tmp/tmp.3poet01eIt + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl logs version-service-5cd4db57-vtrrk + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.1ZMccy00iz + cat /tmp/tmp.3poet01eIt + rm /tmp/tmp.1ZMccy00iz /tmp/tmp.3poet01eIt + return 0 + diff -u /tmp/tmp.IoigtEzcpK/disabled_telemetry.version-service.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/disabled_telemetry.version-service.log.json + [[ disabled == \d\i\s\a\b\l\e\d ]] + [[ recommended == \r\e\c\o\m\m\e\n\d\e\d ]] + diff -u /tmp/tmp.IoigtEzcpK/enabled_telemetry.version-service-cr.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/enabled_telemetry.version-service-cr.log.json + kubectl_bin delete perconapgcluster/minimal ++ mktemp + local LAST_OUT=/tmp/tmp.KBKxpRciy4 ++ mktemp + local LAST_ERR=/tmp/tmp.ntDE6J7u6N + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete perconapgcluster/minimal + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.KBKxpRciy4 perconapgcluster.pg.percona.com "minimal" deleted + cat /tmp/tmp.ntDE6J7u6N + rm /tmp/tmp.KBKxpRciy4 /tmp/tmp.ntDE6J7u6N + return 0 ++ kubectl_bin -n version-service-7914 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OzwR1V6wd4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yD5N9BOEgn ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.OzwR1V6wd4 ++ cat /tmp/tmp.yD5N9BOEgn ++ rm /tmp/tmp.OzwR1V6wd4 /tmp/tmp.yD5N9BOEgn ++ return 0 + wait_job_completion minimal-rmdata-udnr + local job=minimal-rmdata-udnr + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' minimal-rmdata-udnr+ retry=0 ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/minimal-rmdata-udnr -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + kubectl_bin -n version-service-7914 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"false"}}]' ++ mktemp + local LAST_OUT=/tmp/tmp.NgDTgl7NDj ++ mktemp + local LAST_ERR=/tmp/tmp.s2Cv2XDHCY + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"false"}}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.NgDTgl7NDj deployment.apps/postgres-operator patched + cat /tmp/tmp.s2Cv2XDHCY + rm /tmp/tmp.NgDTgl7NDj /tmp/tmp.s2Cv2XDHCY + return 0 + wait_deployment postgres-operator + local name=postgres-operator + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace postgres-operator.... + desc 'Use simple recommended channel' + set +o xtrace ----------------------------------------------------------------------------------- Use simple recommended channel ----------------------------------------------------------------------------------- + cluster=cluster1 + PGO_TAG=1.6.0 + PG_VER=13 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.jBLtzsxzjR ++ mktemp + local LAST_ERR=/tmp/tmp.6htrbSkvuO + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.jBLtzsxzjR secret/cluster1-users-secret created + cat /tmp/tmp.6htrbSkvuO + rm /tmp/tmp.jBLtzsxzjR /tmp/tmp.6htrbSkvuO + return 0 + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.DbYeTR0an0 ++ mktemp + local LAST_ERR=/tmp/tmp.bzjmD4Drz8 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.DbYeTR0an0 secret/cluster1-pmm-secret created + cat /tmp/tmp.bzjmD4Drz8 + rm /tmp/tmp.DbYeTR0an0 /tmp/tmp.bzjmD4Drz8 + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.mnVgGG5SgM ++ mktemp + local LAST_ERR=/tmp/tmp.chHu9CKXwN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.mnVgGG5SgM deployment.apps/pg-client unchanged + cat /tmp/tmp.chHu9CKXwN + rm /tmp/tmp.mnVgGG5SgM /tmp/tmp.chHu9CKXwN + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - spec.database cluster1 + yq w - spec.keepData false + yq w - spec.clustername cluster1 + yq w - metadata.labels.name cluster1 + yq w - spec.keepBackups false + yq w - spec.name cluster1 + yq w - spec.secretsName cluster1-users-secret + yq w - spec.user some-name + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + yq w - metadata.name cluster1 + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + yq w - metadata.labels.crunchy-pgha-scope cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.fNWg6z9xgI ++ mktemp + local LAST_ERR=/tmp/tmp.zZ2p3w5Ova + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.fNWg6z9xgI perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.zZ2p3w5Ova + rm /tmp/tmp.fNWg6z9xgI /tmp/tmp.zZ2p3w5Ova + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1......... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KpJWRUma1D +++ mktemp ++ local LAST_ERR=/tmp/tmp.GSZvMSDeCW ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.KpJWRUma1D ++ cat /tmp/tmp.GSZvMSDeCW ++ rm /tmp/tmp.KpJWRUma1D /tmp/tmp.GSZvMSDeCW ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.eXk89mTSwK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.MplkWUzp4M +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.eXk89mTSwK +++ cat /tmp/tmp.MplkWUzp4M +++ rm /tmp/tmp.eXk89mTSwK /tmp/tmp.MplkWUzp4M +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.m1DRVDZu0m ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ThiKDLFk96 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.m1DRVDZu0m +++ cat /tmp/tmp.ThiKDLFk96 +++ rm /tmp/tmp.m1DRVDZu0m /tmp/tmp.ThiKDLFk96 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tHkKZ8U6Hb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.IElCNvXI8g +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tHkKZ8U6Hb +++ cat /tmp/tmp.IElCNvXI8g +++ rm /tmp/tmp.tHkKZ8U6Hb /tmp/tmp.IElCNvXI8g +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DUaEMwsx5Y ++++ mktemp +++ local LAST_ERR=/tmp/tmp.n3R2EGEMAQ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.DUaEMwsx5Y +++ cat /tmp/tmp.n3R2EGEMAQ +++ rm /tmp/tmp.DUaEMwsx5Y /tmp/tmp.n3R2EGEMAQ +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.i1ehxmjdpu +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.uYpxoek5d1 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.i1ehxmjdpu ++++ cat /tmp/tmp.uYpxoek5d1 ++++ rm /tmp/tmp.i1ehxmjdpu /tmp/tmp.uYpxoek5d1 ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.uXf7cxoIXJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9DmHSOe1a7 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.uXf7cxoIXJ +++ cat /tmp/tmp.9DmHSOe1a7 +++ rm /tmp/tmp.uXf7cxoIXJ /tmp/tmp.9DmHSOe1a7 +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HDarNcdlKb +++ mktemp ++ local LAST_ERR=/tmp/tmp.MzppuX9rNE ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.HDarNcdlKb ++ cat /tmp/tmp.MzppuX9rNE ++ rm /tmp/tmp.HDarNcdlKb /tmp/tmp.MzppuX9rNE ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.IV10CqSxGm ++ mktemp + local LAST_ERR=/tmp/tmp.dc3GxoZIAD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.IV10CqSxGm + cat /tmp/tmp.dc3GxoZIAD + rm /tmp/tmp.IV10CqSxGm /tmp/tmp.dc3GxoZIAD + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZR8rSK3hXL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.DDV6DgIBEi +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZR8rSK3hXL +++ cat /tmp/tmp.DDV6DgIBEi +++ rm /tmp/tmp.ZR8rSK3hXL /tmp/tmp.DDV6DgIBEi +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.xIv2k8JENw ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CtjlVK6CEN +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.xIv2k8JENw +++ cat /tmp/tmp.CtjlVK6CEN +++ rm /tmp/tmp.xIv2k8JENw /tmp/tmp.CtjlVK6CEN +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.EEdVMxugtg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.dsDqEkAr7W +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.EEdVMxugtg +++ cat /tmp/tmp.dsDqEkAr7W +++ rm /tmp/tmp.EEdVMxugtg /tmp/tmp.dsDqEkAr7W +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.HcL7TrlE0w +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.7xLZTj4ss8 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.HcL7TrlE0w ++++ cat /tmp/tmp.7xLZTj4ss8 ++++ rm /tmp/tmp.HcL7TrlE0w /tmp/tmp.7xLZTj4ss8 ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3W5OExjlME ++++ mktemp +++ local LAST_ERR=/tmp/tmp.RoDf0Ske7C +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.3W5OExjlME +++ cat /tmp/tmp.RoDf0Ske7C +++ rm /tmp/tmp.3W5OExjlME /tmp/tmp.RoDf0Ske7C +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4n0YLxoz08 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gV2WLMpLsZ ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.4n0YLxoz08 ++ cat /tmp/tmp.gV2WLMpLsZ ++ rm /tmp/tmp.4n0YLxoz08 /tmp/tmp.gV2WLMpLsZ ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.UZWMJrQCVe ++ mktemp + local LAST_ERR=/tmp/tmp.ZZK1LtIHXF + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.UZWMJrQCVe + cat /tmp/tmp.ZZK1LtIHXF + rm /tmp/tmp.UZWMJrQCVe /tmp/tmp.ZZK1LtIHXF + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.vPUHa6qd5P ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9KaRiLiDj5 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.vPUHa6qd5P +++ cat /tmp/tmp.9KaRiLiDj5 +++ rm /tmp/tmp.vPUHa6qd5P /tmp/tmp.9KaRiLiDj5 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VVfPAPRN6q ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LbzvwlUbiQ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.VVfPAPRN6q +++ cat /tmp/tmp.LbzvwlUbiQ +++ rm /tmp/tmp.VVfPAPRN6q /tmp/tmp.LbzvwlUbiQ +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wOy3eDnoiE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1DcnbMK9hu +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.wOy3eDnoiE +++ cat /tmp/tmp.1DcnbMK9hu +++ rm /tmp/tmp.wOy3eDnoiE /tmp/tmp.1DcnbMK9hu +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.fHe9rsm8ME +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.O0BOYsgLBO ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.fHe9rsm8ME ++++ cat /tmp/tmp.O0BOYsgLBO ++++ rm /tmp/tmp.fHe9rsm8ME /tmp/tmp.O0BOYsgLBO ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mJhnoESZ3s ++++ mktemp +++ local LAST_ERR=/tmp/tmp.3yZKXn1WxB +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.mJhnoESZ3s +++ cat /tmp/tmp.3yZKXn1WxB +++ rm /tmp/tmp.mJhnoESZ3s /tmp/tmp.3yZKXn1WxB +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TEIuIMq1Rf +++ mktemp ++ local LAST_ERR=/tmp/tmp.lIHi6GTn1H ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.TEIuIMq1Rf ++ cat /tmp/tmp.lIHi6GTn1H ++ rm /tmp/tmp.TEIuIMq1Rf /tmp/tmp.lIHi6GTn1H ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.NfZIwOStlv ++ mktemp + local LAST_ERR=/tmp/tmp.xjezYcyVvA + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.NfZIwOStlv + cat /tmp/tmp.xjezYcyVvA + rm /tmp/tmp.NfZIwOStlv /tmp/tmp.xjezYcyVvA + return 0 + sleep 10 + apply_version cluster1 9.9.9 recommended 13 http://version-service:11000 0 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=recommended + local db_version=13 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=0 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.nuvj1AAeu1 ++ mktemp + local LAST_ERR=/tmp/tmp.lvQGH4wlN1 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nuvj1AAeu1 perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.lvQGH4wlN1 + rm /tmp/tmp.nuvj1AAeu1 /tmp/tmp.lvQGH4wlN1 + return 0 + sleep 10 + wait_smart_update cluster1 0 + local cluster=cluster1 + local target=0 + set +o xtrace ......................................................................... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.Jz4tHmQ5iF ++ mktemp + local LAST_ERR=/tmp/tmp.3XHbrK9CLU + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Jz4tHmQ5iF job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.3XHbrK9CLU + rm /tmp/tmp.Jz4tHmQ5iF /tmp/tmp.3XHbrK9CLU + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + kubectl_bin apply -f - + yq w - spec.parameters.pg-cluster cluster1 + yq w - spec.namespace version-service-7914 ++ mktemp ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' + yq w - spec.parameters.job-name fulldata-cluster1 +++ mktemp + local LAST_OUT=/tmp/tmp.CwGIFo880g ++ mktemp ++ local LAST_OUT=/tmp/tmp.2anRowARyV +++ mktemp + local LAST_ERR=/tmp/tmp.Ts3Gb8Z24S + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ local LAST_ERR=/tmp/tmp.BOYdq9CgyZ ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + for i in '$(seq 0 2)' +++ seq 0 2 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.2anRowARyV ++ cat /tmp/tmp.BOYdq9CgyZ ++ rm /tmp/tmp.2anRowARyV /tmp/tmp.BOYdq9CgyZ ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-57f668d478-59n6c + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.CwGIFo880g pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.Ts3Gb8Z24S + rm /tmp/tmp.CwGIFo880g /tmp/tmp.Ts3Gb8Z24S + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 recommended 13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=recommended + local db_version=13 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.X8hBiywWCK ++ mktemp + local LAST_ERR=/tmp/tmp.oDJhNzu7uC + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.X8hBiywWCK + cat /tmp/tmp.oDJhNzu7uC + rm /tmp/tmp.X8hBiywWCK /tmp/tmp.oDJhNzu7uC + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Use major version recommended channel' + set +o xtrace ----------------------------------------------------------------------------------- Use major version recommended channel ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + kubectl_bin -n version-service-7914 apply -f - + yq w - stringData.some-name some-name_pass ++ mktemp + local LAST_OUT=/tmp/tmp.HmidgsGqau ++ mktemp + local LAST_ERR=/tmp/tmp.kwp8bFrZiW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.HmidgsGqau secret/cluster1-users-secret created + cat /tmp/tmp.kwp8bFrZiW + rm /tmp/tmp.HmidgsGqau /tmp/tmp.kwp8bFrZiW + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.t5bmiqo8qf ++ mktemp + local LAST_ERR=/tmp/tmp.0dyqiedT2w + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.t5bmiqo8qf secret/cluster1-pmm-secret configured + cat /tmp/tmp.0dyqiedT2w + rm /tmp/tmp.t5bmiqo8qf /tmp/tmp.0dyqiedT2w + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.z8kHWJDXvN ++ mktemp + local LAST_ERR=/tmp/tmp.R3OL8BHSER + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.z8kHWJDXvN deployment.apps/pg-client unchanged + cat /tmp/tmp.R3OL8BHSER + rm /tmp/tmp.z8kHWJDXvN /tmp/tmp.R3OL8BHSER + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - spec.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.keepData false + yq w - spec.clustername cluster1 + yq w - spec.user some-name + yq w - metadata.name cluster1 + yq w - spec.database cluster1 + yq w - spec.keepBackups false + yq w - spec.secretsName cluster1-users-secret + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + yq w - metadata.labels.name cluster1 + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.7A7B17DCWX ++ mktemp + local LAST_ERR=/tmp/tmp.5Erhjc444X + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.7A7B17DCWX perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.5Erhjc444X + rm /tmp/tmp.7A7B17DCWX /tmp/tmp.5Erhjc444X + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1........ + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Q0gkEUtBBE +++ mktemp ++ local LAST_ERR=/tmp/tmp.fjHpoMTn6i ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.Q0gkEUtBBE ++ cat /tmp/tmp.fjHpoMTn6i ++ rm /tmp/tmp.Q0gkEUtBBE /tmp/tmp.fjHpoMTn6i ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.kpBmEyYjxY ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NArNUFNdss +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.kpBmEyYjxY +++ cat /tmp/tmp.NArNUFNdss +++ rm /tmp/tmp.kpBmEyYjxY /tmp/tmp.NArNUFNdss +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.QeNva6C8PZ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.paKP63hLoE +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.QeNva6C8PZ +++ cat /tmp/tmp.paKP63hLoE +++ rm /tmp/tmp.QeNva6C8PZ /tmp/tmp.paKP63hLoE +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WZia02APG4 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.FHFVwbyI0Z +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.WZia02APG4 +++ cat /tmp/tmp.FHFVwbyI0Z +++ rm /tmp/tmp.WZia02APG4 /tmp/tmp.FHFVwbyI0Z +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZVjx2JYXAh ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LKdMcFokOO +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZVjx2JYXAh +++ cat /tmp/tmp.LKdMcFokOO +++ rm /tmp/tmp.ZVjx2JYXAh /tmp/tmp.LKdMcFokOO +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.TdNBFXaa94 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.UqutGsEaea ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.TdNBFXaa94 ++++ cat /tmp/tmp.UqutGsEaea ++++ rm /tmp/tmp.TdNBFXaa94 /tmp/tmp.UqutGsEaea ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rKSPv0KFGk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XDEZZfvH9X +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rKSPv0KFGk +++ cat /tmp/tmp.XDEZZfvH9X +++ rm /tmp/tmp.rKSPv0KFGk /tmp/tmp.XDEZZfvH9X +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o9R8xaXjsj +++ mktemp ++ local LAST_ERR=/tmp/tmp.hEbM4UMQOA ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.o9R8xaXjsj ++ cat /tmp/tmp.hEbM4UMQOA ++ rm /tmp/tmp.o9R8xaXjsj /tmp/tmp.hEbM4UMQOA ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.2iyyTsQnoC ++ mktemp + local LAST_ERR=/tmp/tmp.ereXn3DJPV + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.2iyyTsQnoC + cat /tmp/tmp.ereXn3DJPV + rm /tmp/tmp.2iyyTsQnoC /tmp/tmp.ereXn3DJPV + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.cg71xF3gpe ++++ mktemp +++ local LAST_ERR=/tmp/tmp.w8th6t12X7 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.cg71xF3gpe +++ cat /tmp/tmp.w8th6t12X7 +++ rm /tmp/tmp.cg71xF3gpe /tmp/tmp.w8th6t12X7 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1krMiaeUTd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vIzh7fo0Ua +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.1krMiaeUTd +++ cat /tmp/tmp.vIzh7fo0Ua +++ rm /tmp/tmp.1krMiaeUTd /tmp/tmp.vIzh7fo0Ua +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wEQB87wEeU ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NCCN5cOt18 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.wEQB87wEeU +++ cat /tmp/tmp.NCCN5cOt18 +++ rm /tmp/tmp.wEQB87wEeU /tmp/tmp.NCCN5cOt18 +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.zYSHvQL6iQ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.nYRshIsxmM ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.zYSHvQL6iQ ++++ cat /tmp/tmp.nYRshIsxmM ++++ rm /tmp/tmp.zYSHvQL6iQ /tmp/tmp.nYRshIsxmM ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.PuHvnLCS4H ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ryFmASipKS +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.PuHvnLCS4H +++ cat /tmp/tmp.ryFmASipKS +++ rm /tmp/tmp.PuHvnLCS4H /tmp/tmp.ryFmASipKS +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bbFbtODxMN +++ mktemp ++ local LAST_ERR=/tmp/tmp.dURDc1nfUf ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.bbFbtODxMN ++ cat /tmp/tmp.dURDc1nfUf ++ rm /tmp/tmp.bbFbtODxMN /tmp/tmp.dURDc1nfUf ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.PopximkTDt ++ mktemp + local LAST_ERR=/tmp/tmp.Rg092yxpKE + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PopximkTDt + cat /tmp/tmp.Rg092yxpKE + rm /tmp/tmp.PopximkTDt /tmp/tmp.Rg092yxpKE + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qxEuR95fSF ++++ mktemp +++ local LAST_ERR=/tmp/tmp.8sEDsyHl7g +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qxEuR95fSF +++ cat /tmp/tmp.8sEDsyHl7g +++ rm /tmp/tmp.qxEuR95fSF /tmp/tmp.8sEDsyHl7g +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.f6DD0WiwXj ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bJUUsMUdxo +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.f6DD0WiwXj +++ cat /tmp/tmp.bJUUsMUdxo +++ rm /tmp/tmp.f6DD0WiwXj /tmp/tmp.bJUUsMUdxo +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qNalYOPCxo ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VlORFyLd4Q +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qNalYOPCxo +++ cat /tmp/tmp.VlORFyLd4Q +++ rm /tmp/tmp.qNalYOPCxo /tmp/tmp.VlORFyLd4Q +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.c5gaR7Qsu5 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.IBexqj6Tfd ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.c5gaR7Qsu5 ++++ cat /tmp/tmp.IBexqj6Tfd ++++ rm /tmp/tmp.c5gaR7Qsu5 /tmp/tmp.IBexqj6Tfd ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hg4Tzn5OQv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.RG1CTlmhNt +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.hg4Tzn5OQv +++ cat /tmp/tmp.RG1CTlmhNt +++ rm /tmp/tmp.hg4Tzn5OQv /tmp/tmp.RG1CTlmhNt +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MR3osoYvVB +++ mktemp ++ local LAST_ERR=/tmp/tmp.w0XFpWATWA ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.MR3osoYvVB ++ cat /tmp/tmp.w0XFpWATWA ++ rm /tmp/tmp.MR3osoYvVB /tmp/tmp.w0XFpWATWA ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.FZ3UQLFoDn ++ mktemp + local LAST_ERR=/tmp/tmp.9ctAjtPsRc + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FZ3UQLFoDn + cat /tmp/tmp.9ctAjtPsRc + rm /tmp/tmp.FZ3UQLFoDn /tmp/tmp.9ctAjtPsRc + return 0 + sleep 10 + apply_version cluster1 9.9.9 12-recommended 12 http://version-service:11000 1 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-recommended + local db_version=12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=1 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.OuXqC4blCv ++ mktemp + local LAST_ERR=/tmp/tmp.JWe0jIJi3r + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.OuXqC4blCv perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.JWe0jIJi3r + rm /tmp/tmp.OuXqC4blCv /tmp/tmp.JWe0jIJi3r + return 0 + sleep 10 + wait_smart_update cluster1 1 + local cluster=cluster1 + local target=1 + set +o xtrace ............................................................... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.VT22V37Y94 ++ mktemp + local LAST_ERR=/tmp/tmp.Nv4GwadCco + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.VT22V37Y94 job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.Nv4GwadCco + rm /tmp/tmp.VT22V37Y94 /tmp/tmp.Nv4GwadCco + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - spec.namespace version-service-7914 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + kubectl_bin apply -f - ++ mktemp ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' + yq w - spec.parameters.pg-cluster cluster1 + local LAST_OUT=/tmp/tmp.piVP2iBADD +++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.7RVl4rb3nZ + local exit_status=0 + local timeout=4 + set +o errexit ++ local LAST_OUT=/tmp/tmp.f7HyUIUGuB ++ seq 0 2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DpjnOey4We ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.f7HyUIUGuB ++ cat /tmp/tmp.DpjnOey4We ++ rm /tmp/tmp.f7HyUIUGuB /tmp/tmp.DpjnOey4We ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-5f647d77b5-dxbk9 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.piVP2iBADD pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.7RVl4rb3nZ + rm /tmp/tmp.piVP2iBADD /tmp/tmp.7RVl4rb3nZ + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 12-recommended 12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-recommended + local db_version=12 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.xRaHqEb3hA ++ mktemp + local LAST_ERR=/tmp/tmp.AvIF50JxGG + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xRaHqEb3hA + cat /tmp/tmp.AvIF50JxGG + rm /tmp/tmp.xRaHqEb3hA /tmp/tmp.AvIF50JxGG + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-12-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ....Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Updating PG cluster with 14 from 13' + set +o xtrace ----------------------------------------------------------------------------------- Updating PG cluster with 14 from 13 ----------------------------------------------------------------------------------- + PG_VER=13 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.czo8MsGnVw ++ mktemp + local LAST_ERR=/tmp/tmp.3xHqIefsQv + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.czo8MsGnVw secret/cluster1-users-secret created + cat /tmp/tmp.3xHqIefsQv + rm /tmp/tmp.czo8MsGnVw /tmp/tmp.3xHqIefsQv + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.tFpYBOr5qC ++ mktemp + local LAST_ERR=/tmp/tmp.NMd3BzSSbO + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + yq w - metadata.name cluster1-pmm-secret + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.tFpYBOr5qC secret/cluster1-pmm-secret configured + cat /tmp/tmp.NMd3BzSSbO + rm /tmp/tmp.tFpYBOr5qC /tmp/tmp.NMd3BzSSbO + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.kbvRYvxZfG ++ mktemp + local LAST_ERR=/tmp/tmp.EbDP7NDO9K + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.kbvRYvxZfG deployment.apps/pg-client unchanged + cat /tmp/tmp.EbDP7NDO9K + rm /tmp/tmp.kbvRYvxZfG /tmp/tmp.EbDP7NDO9K + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.name cluster1 + yq w - spec.clustername cluster1 + yq w - spec.database cluster1 + yq w - spec.user some-name + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + yq w - spec.keepBackups false + yq w - spec.name cluster1 + yq w - spec.keepData false + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.labels.deployment-name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.gG1GRVh1ON ++ mktemp + local LAST_ERR=/tmp/tmp.iRNjfHVxxt + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gG1GRVh1ON perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.iRNjfHVxxt + rm /tmp/tmp.gG1GRVh1ON /tmp/tmp.iRNjfHVxxt + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1........... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dVkj5ogvaV +++ mktemp ++ local LAST_ERR=/tmp/tmp.ofMpBQb33Z ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.dVkj5ogvaV ++ cat /tmp/tmp.ofMpBQb33Z ++ rm /tmp/tmp.dVkj5ogvaV /tmp/tmp.ofMpBQb33Z ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Av8zYrJxxs ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NeGwElU3y4 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Av8zYrJxxs +++ cat /tmp/tmp.NeGwElU3y4 +++ rm /tmp/tmp.Av8zYrJxxs /tmp/tmp.NeGwElU3y4 +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3Jk0g7aJ5z ++++ mktemp +++ local LAST_ERR=/tmp/tmp.o08sFeUsKV +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.3Jk0g7aJ5z +++ cat /tmp/tmp.o08sFeUsKV +++ rm /tmp/tmp.3Jk0g7aJ5z /tmp/tmp.o08sFeUsKV +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.TExlLiEDWp ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bvvwKObqRs +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.TExlLiEDWp +++ cat /tmp/tmp.bvvwKObqRs +++ rm /tmp/tmp.TExlLiEDWp /tmp/tmp.bvvwKObqRs +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lp2tIgDZYx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.HE89WGMqjR +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.lp2tIgDZYx +++ cat /tmp/tmp.HE89WGMqjR +++ rm /tmp/tmp.lp2tIgDZYx /tmp/tmp.HE89WGMqjR +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.yDgcTXsMmR +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.8osPr91pGW ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.yDgcTXsMmR ++++ cat /tmp/tmp.8osPr91pGW ++++ rm /tmp/tmp.yDgcTXsMmR /tmp/tmp.8osPr91pGW ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tnXyBrN7Hx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.qdF8fWBz2e +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tnXyBrN7Hx +++ cat /tmp/tmp.qdF8fWBz2e +++ rm /tmp/tmp.tnXyBrN7Hx /tmp/tmp.qdF8fWBz2e +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XpGlUbmCmH +++ mktemp ++ local LAST_ERR=/tmp/tmp.rSOVu2uxy1 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.XpGlUbmCmH ++ cat /tmp/tmp.rSOVu2uxy1 ++ rm /tmp/tmp.XpGlUbmCmH /tmp/tmp.rSOVu2uxy1 ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.6kPhmWuz1i ++ mktemp + local LAST_ERR=/tmp/tmp.H2dOGfZdFw + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.6kPhmWuz1i + cat /tmp/tmp.H2dOGfZdFw + rm /tmp/tmp.6kPhmWuz1i /tmp/tmp.H2dOGfZdFw + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.IkghascQO0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5PMZqnzjVF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.IkghascQO0 +++ cat /tmp/tmp.5PMZqnzjVF +++ rm /tmp/tmp.IkghascQO0 /tmp/tmp.5PMZqnzjVF +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Gn9KopOjn3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fNtbFPp5pn +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Gn9KopOjn3 +++ cat /tmp/tmp.fNtbFPp5pn +++ rm /tmp/tmp.Gn9KopOjn3 /tmp/tmp.fNtbFPp5pn +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.HIU33MQld6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.AdC4YygBi5 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.HIU33MQld6 +++ cat /tmp/tmp.AdC4YygBi5 +++ rm /tmp/tmp.HIU33MQld6 /tmp/tmp.AdC4YygBi5 +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.rmDu0WYayk +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.JnYcMfYjie ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.rmDu0WYayk ++++ cat /tmp/tmp.JnYcMfYjie ++++ rm /tmp/tmp.rmDu0WYayk /tmp/tmp.JnYcMfYjie ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6cgsOjjo2L ++++ mktemp +++ local LAST_ERR=/tmp/tmp.6hEkAPXKMN +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.6cgsOjjo2L +++ cat /tmp/tmp.6hEkAPXKMN +++ rm /tmp/tmp.6cgsOjjo2L /tmp/tmp.6hEkAPXKMN +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kySquRMerJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.pW7wmB51Sy ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.kySquRMerJ ++ cat /tmp/tmp.pW7wmB51Sy ++ rm /tmp/tmp.kySquRMerJ /tmp/tmp.pW7wmB51Sy ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.U2GjkmlYO3 ++ mktemp + local LAST_ERR=/tmp/tmp.1LOpCyfaQg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.U2GjkmlYO3 + cat /tmp/tmp.1LOpCyfaQg + rm /tmp/tmp.U2GjkmlYO3 /tmp/tmp.1LOpCyfaQg + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3vUu3xClU6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yQOuaIeCvv +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.3vUu3xClU6 +++ cat /tmp/tmp.yQOuaIeCvv +++ rm /tmp/tmp.3vUu3xClU6 /tmp/tmp.yQOuaIeCvv +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.PnzGcIBtwN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.HnjAHUXVYD +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.PnzGcIBtwN +++ cat /tmp/tmp.HnjAHUXVYD +++ rm /tmp/tmp.PnzGcIBtwN /tmp/tmp.HnjAHUXVYD +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MJnJQcPNZ8 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.sJ3aG3I1WR +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MJnJQcPNZ8 +++ cat /tmp/tmp.sJ3aG3I1WR +++ rm /tmp/tmp.MJnJQcPNZ8 /tmp/tmp.sJ3aG3I1WR +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.T6z3kh6flU +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.pWTKrfz8Mp ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.T6z3kh6flU ++++ cat /tmp/tmp.pWTKrfz8Mp ++++ rm /tmp/tmp.T6z3kh6flU /tmp/tmp.pWTKrfz8Mp ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.H7mM72kVQT ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CzZCkHJhKV +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.H7mM72kVQT +++ cat /tmp/tmp.CzZCkHJhKV +++ rm /tmp/tmp.H7mM72kVQT /tmp/tmp.CzZCkHJhKV +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JzLQbhHvqZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.GQskvcVpPx ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.JzLQbhHvqZ ++ cat /tmp/tmp.GQskvcVpPx ++ rm /tmp/tmp.JzLQbhHvqZ /tmp/tmp.GQskvcVpPx ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.iBnrIcZw6f ++ mktemp + local LAST_ERR=/tmp/tmp.lg9pMX30b5 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.iBnrIcZw6f + cat /tmp/tmp.lg9pMX30b5 + rm /tmp/tmp.iBnrIcZw6f /tmp/tmp.lg9pMX30b5 + return 0 + sleep 10 + apply_version cluster1 9.9.9 14-recommended 1.6.0-13 http://version-service:11000 1 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=14-recommended + local db_version=1.6.0-13 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=1 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"14-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.MwVqVa59z1 ++ mktemp + local LAST_ERR=/tmp/tmp.xkQCoKE2Kg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"14-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.MwVqVa59z1 perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.xkQCoKE2Kg + rm /tmp/tmp.MwVqVa59z1 /tmp/tmp.xkQCoKE2Kg + return 0 + sleep 10 + wait_smart_update cluster1 1 + local cluster=cluster1 + local target=1 + set +o xtrace + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1................ + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.JYMOSoxJcJ ++ mktemp + local LAST_ERR=/tmp/tmp.KjHEXUQDfF + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.JYMOSoxJcJ job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.KjHEXUQDfF + rm /tmp/tmp.JYMOSoxJcJ /tmp/tmp.KjHEXUQDfF + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + kubectl_bin apply -f - + yq w - spec.namespace version-service-7914 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w - spec.parameters.pg-cluster cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full ++ mktemp ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' +++ mktemp + local LAST_OUT=/tmp/tmp.LWvySnYUqY ++ mktemp + local LAST_ERR=/tmp/tmp.pqAsUuEq3o + local exit_status=0 + local timeout=4 + set +o errexit ++ local LAST_OUT=/tmp/tmp.7F1eJz8R0H ++ seq 0 2 +++ mktemp + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - ++ local LAST_ERR=/tmp/tmp.2EniBEHmFd ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.7F1eJz8R0H ++ cat /tmp/tmp.2EniBEHmFd ++ rm /tmp/tmp.7F1eJz8R0H /tmp/tmp.2EniBEHmFd ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-9f8cd5549-g7dzf + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.LWvySnYUqY pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.pqAsUuEq3o + rm /tmp/tmp.LWvySnYUqY /tmp/tmp.pqAsUuEq3o + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 14-recommended 1.6.0-13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=14-recommended + local db_version=1.6.0-13 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.jTa1dYXYVr ++ mktemp + local LAST_ERR=/tmp/tmp.gjhe7M54Wx + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.jTa1dYXYVr + cat /tmp/tmp.gjhe7M54Wx + rm /tmp/tmp.jTa1dYXYVr /tmp/tmp.gjhe7M54Wx + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-14-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/1.6.0-13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ..Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster update with the latest image by version service' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster update with the latest image by version service ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.SuQbLy12xA ++ mktemp + local LAST_ERR=/tmp/tmp.Fu4owlCVEE + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.SuQbLy12xA secret/cluster1-users-secret created + cat /tmp/tmp.Fu4owlCVEE + rm /tmp/tmp.SuQbLy12xA /tmp/tmp.Fu4owlCVEE + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ueX1CoRnJl ++ mktemp + local LAST_ERR=/tmp/tmp.8cWiYgBTVE + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ueX1CoRnJl secret/cluster1-pmm-secret configured + cat /tmp/tmp.8cWiYgBTVE + rm /tmp/tmp.ueX1CoRnJl /tmp/tmp.8cWiYgBTVE + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.lwagwCSwpH ++ mktemp + local LAST_ERR=/tmp/tmp.vxAsCxjeuN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.lwagwCSwpH deployment.apps/pg-client unchanged + cat /tmp/tmp.vxAsCxjeuN + rm /tmp/tmp.lwagwCSwpH /tmp/tmp.vxAsCxjeuN + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.name cluster1 + yq w - spec.name cluster1 + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + yq w - spec.user some-name + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + yq w - spec.keepBackups false + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + yq w - spec.keepData false + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - spec.clustername cluster1 + yq w - spec.database cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.98XbamNVIa ++ mktemp + local LAST_ERR=/tmp/tmp.MkpVnaCUpH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.98XbamNVIa perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.MkpVnaCUpH + rm /tmp/tmp.98XbamNVIa /tmp/tmp.MkpVnaCUpH + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1......... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dT70Feo8NH +++ mktemp ++ local LAST_ERR=/tmp/tmp.dvxJLSyZPl ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.dT70Feo8NH ++ cat /tmp/tmp.dvxJLSyZPl ++ rm /tmp/tmp.dT70Feo8NH /tmp/tmp.dvxJLSyZPl ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.AbHtMv5gT8 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ETmmndGfGw +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.AbHtMv5gT8 +++ cat /tmp/tmp.ETmmndGfGw +++ rm /tmp/tmp.AbHtMv5gT8 /tmp/tmp.ETmmndGfGw +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.gjtPCVads4 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.KRDlh2RAcb +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.gjtPCVads4 +++ cat /tmp/tmp.KRDlh2RAcb +++ rm /tmp/tmp.gjtPCVads4 /tmp/tmp.KRDlh2RAcb +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.G8Ws6n1EXS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.u5MKnWIltX +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.G8Ws6n1EXS +++ cat /tmp/tmp.u5MKnWIltX +++ rm /tmp/tmp.G8Ws6n1EXS /tmp/tmp.u5MKnWIltX +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.7u1zpAqwRC ++++ mktemp +++ local LAST_ERR=/tmp/tmp.hxr8hluFUh +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.7u1zpAqwRC +++ cat /tmp/tmp.hxr8hluFUh +++ rm /tmp/tmp.7u1zpAqwRC /tmp/tmp.hxr8hluFUh +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.PaoGbPKTOp +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.O4qRjw2qqg ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.PaoGbPKTOp ++++ cat /tmp/tmp.O4qRjw2qqg ++++ rm /tmp/tmp.PaoGbPKTOp /tmp/tmp.O4qRjw2qqg ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wI7Li8isUt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.mTSk1YRmjs +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.wI7Li8isUt +++ cat /tmp/tmp.mTSk1YRmjs +++ rm /tmp/tmp.wI7Li8isUt /tmp/tmp.mTSk1YRmjs +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5wNNXuBVb7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.BOVwOJab9E ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.5wNNXuBVb7 ++ cat /tmp/tmp.BOVwOJab9E ++ rm /tmp/tmp.5wNNXuBVb7 /tmp/tmp.BOVwOJab9E ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.Mou2TaM2uB ++ mktemp + local LAST_ERR=/tmp/tmp.j2aDynnjr0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Mou2TaM2uB + cat /tmp/tmp.j2aDynnjr0 + rm /tmp/tmp.Mou2TaM2uB /tmp/tmp.j2aDynnjr0 + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.sa5eaue6nK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VJlgzLRN7N +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.sa5eaue6nK +++ cat /tmp/tmp.VJlgzLRN7N +++ rm /tmp/tmp.sa5eaue6nK /tmp/tmp.VJlgzLRN7N +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.NYA4Su7T4C ++++ mktemp +++ local LAST_ERR=/tmp/tmp.EJW5zxJ7Xi +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.NYA4Su7T4C +++ cat /tmp/tmp.EJW5zxJ7Xi +++ rm /tmp/tmp.NYA4Su7T4C /tmp/tmp.EJW5zxJ7Xi +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SNYP0m1uw8 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ux6vnO2IiI +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.SNYP0m1uw8 +++ cat /tmp/tmp.ux6vnO2IiI +++ rm /tmp/tmp.SNYP0m1uw8 /tmp/tmp.ux6vnO2IiI +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.5fzJFwa0Hc +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.tXcnu4TxVt ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.5fzJFwa0Hc ++++ cat /tmp/tmp.tXcnu4TxVt ++++ rm /tmp/tmp.5fzJFwa0Hc /tmp/tmp.tXcnu4TxVt ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.LNOt2FiqDA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.4J8h3JhYxh +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.LNOt2FiqDA +++ cat /tmp/tmp.4J8h3JhYxh +++ rm /tmp/tmp.LNOt2FiqDA /tmp/tmp.4J8h3JhYxh +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f0k79Ont3t +++ mktemp ++ local LAST_ERR=/tmp/tmp.8etyzVzhJ2 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.f0k79Ont3t ++ cat /tmp/tmp.8etyzVzhJ2 ++ rm /tmp/tmp.f0k79Ont3t /tmp/tmp.8etyzVzhJ2 ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.LJFoW2lAE8 ++ mktemp + local LAST_ERR=/tmp/tmp.yQXBbo7PGu + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.LJFoW2lAE8 + cat /tmp/tmp.yQXBbo7PGu + rm /tmp/tmp.LJFoW2lAE8 /tmp/tmp.yQXBbo7PGu + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rch41Vuwsr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1dsSnQ1Wiy +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rch41Vuwsr +++ cat /tmp/tmp.1dsSnQ1Wiy +++ rm /tmp/tmp.rch41Vuwsr /tmp/tmp.1dsSnQ1Wiy +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.KvMKneek6M ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CBhbVgjwVo +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.KvMKneek6M +++ cat /tmp/tmp.CBhbVgjwVo +++ rm /tmp/tmp.KvMKneek6M /tmp/tmp.CBhbVgjwVo +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.iSMZoYhcSG ++++ mktemp +++ local LAST_ERR=/tmp/tmp.SufvIF7pZa +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.iSMZoYhcSG +++ cat /tmp/tmp.SufvIF7pZa +++ rm /tmp/tmp.iSMZoYhcSG /tmp/tmp.SufvIF7pZa +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Ci4SBmNkAi +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.keJ8oCkoQM ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.Ci4SBmNkAi ++++ cat /tmp/tmp.keJ8oCkoQM ++++ rm /tmp/tmp.Ci4SBmNkAi /tmp/tmp.keJ8oCkoQM ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.TBdPcmfW0S ++++ mktemp +++ local LAST_ERR=/tmp/tmp.P1iF33Hysr +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.TBdPcmfW0S +++ cat /tmp/tmp.P1iF33Hysr +++ rm /tmp/tmp.TBdPcmfW0S /tmp/tmp.P1iF33Hysr +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EHFlc2UXFL +++ mktemp ++ local LAST_ERR=/tmp/tmp.VSoAKThARE ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.EHFlc2UXFL ++ cat /tmp/tmp.VSoAKThARE ++ rm /tmp/tmp.EHFlc2UXFL /tmp/tmp.VSoAKThARE ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.TsimbYFYpI ++ mktemp + local LAST_ERR=/tmp/tmp.HKcTfAgZ36 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.TsimbYFYpI + cat /tmp/tmp.HKcTfAgZ36 + rm /tmp/tmp.TsimbYFYpI /tmp/tmp.HKcTfAgZ36 + return 0 + sleep 10 + apply_version cluster1 9.9.9 latest 12 http://version-service:11000 2 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=2 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.5efrITjyZr ++ mktemp + local LAST_ERR=/tmp/tmp.265hSU9Ul2 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.5efrITjyZr perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.265hSU9Ul2 + rm /tmp/tmp.5efrITjyZr /tmp/tmp.265hSU9Ul2 + return 0 + sleep 10 + wait_smart_update cluster1 2 + local cluster=cluster1 + local target=2 + set +o xtrace ........................................................ + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.KcjJsUYfot ++ mktemp + local LAST_ERR=/tmp/tmp.2Sk7rP6iFU + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.KcjJsUYfot job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.2Sk7rP6iFU + rm /tmp/tmp.KcjJsUYfot /tmp/tmp.2Sk7rP6iFU + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.namespace version-service-7914 + yq w - spec.parameters.pg-cluster cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' + kubectl_bin apply -f - + yq w - metadata.name fulldata-cluster1 +++ mktemp ++ local LAST_OUT=/tmp/tmp.f8Dpawv0hh ++ mktemp +++ mktemp + local LAST_OUT=/tmp/tmp.w8HcyawXWi ++ mktemp ++ local LAST_ERR=/tmp/tmp.KBPoLCgZab ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + local LAST_ERR=/tmp/tmp.4NDKMn3lhy + local exit_status=0 + local timeout=4 + set +o errexit +++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.f8Dpawv0hh ++ cat /tmp/tmp.KBPoLCgZab ++ rm /tmp/tmp.f8Dpawv0hh /tmp/tmp.KBPoLCgZab ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-5f647d77b5-9t6dj + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.w8HcyawXWi pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.4NDKMn3lhy + rm /tmp/tmp.w8HcyawXWi /tmp/tmp.4NDKMn3lhy + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 latest 12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=12 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.QdHPU9SdFu ++ mktemp + local LAST_ERR=/tmp/tmp.ffimgT37gG + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.QdHPU9SdFu + cat /tmp/tmp.ffimgT37gG + rm /tmp/tmp.QdHPU9SdFu /tmp/tmp.ffimgT37gG + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-latest-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster with major version latest release channel' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster with major version latest release channel ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - + yq w - metadata.name cluster1-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.nDKxu9Bcc1 ++ mktemp + local LAST_ERR=/tmp/tmp.xPstkwqUD3 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nDKxu9Bcc1 secret/cluster1-users-secret created + cat /tmp/tmp.xPstkwqUD3 + rm /tmp/tmp.nDKxu9Bcc1 /tmp/tmp.xPstkwqUD3 + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ayTWio9bSE ++ mktemp + local LAST_ERR=/tmp/tmp.ESWyjxVIYo + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ayTWio9bSE secret/cluster1-pmm-secret configured + cat /tmp/tmp.ESWyjxVIYo + rm /tmp/tmp.ayTWio9bSE /tmp/tmp.ESWyjxVIYo + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.pu0ximsOB8 ++ mktemp + local LAST_ERR=/tmp/tmp.TGIDAkGTdW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.pu0ximsOB8 deployment.apps/pg-client unchanged + cat /tmp/tmp.TGIDAkGTdW + rm /tmp/tmp.pu0ximsOB8 /tmp/tmp.TGIDAkGTdW + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.name cluster1 + yq w - spec.keepData false + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.clustername cluster1 + yq w - spec.keepBackups false + yq w - spec.user some-name + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + yq w - spec.secretsName cluster1-users-secret + yq w - spec.name cluster1 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + yq w - spec.database cluster1 + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.Y0TocSe8dt ++ mktemp + local LAST_ERR=/tmp/tmp.zLLXdOFnQQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Y0TocSe8dt perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.zLLXdOFnQQ + rm /tmp/tmp.Y0TocSe8dt /tmp/tmp.zLLXdOFnQQ + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1....... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tbPiGuJYST +++ mktemp ++ local LAST_ERR=/tmp/tmp.Xkm0yZGF7Y ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.tbPiGuJYST ++ cat /tmp/tmp.Xkm0yZGF7Y ++ rm /tmp/tmp.tbPiGuJYST /tmp/tmp.Xkm0yZGF7Y ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.zCzmLZ8O6a ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1Rk6vo4Ldj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.zCzmLZ8O6a +++ cat /tmp/tmp.1Rk6vo4Ldj +++ rm /tmp/tmp.zCzmLZ8O6a /tmp/tmp.1Rk6vo4Ldj +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rOnnSm6aqY ++++ mktemp +++ local LAST_ERR=/tmp/tmp.neQ6M28Z6v +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rOnnSm6aqY +++ cat /tmp/tmp.neQ6M28Z6v +++ rm /tmp/tmp.rOnnSm6aqY /tmp/tmp.neQ6M28Z6v +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.XkhRxV8gbk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5SMUFGEMU3 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.XkhRxV8gbk +++ cat /tmp/tmp.5SMUFGEMU3 +++ rm /tmp/tmp.XkhRxV8gbk /tmp/tmp.5SMUFGEMU3 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.BkwjC3f7Mr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.GipXDe3Bwx +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.BkwjC3f7Mr +++ cat /tmp/tmp.GipXDe3Bwx +++ rm /tmp/tmp.BkwjC3f7Mr /tmp/tmp.GipXDe3Bwx +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.2EH9LJehGa ++++ mktemp +++ local LAST_ERR=/tmp/tmp.eZd8tzkM9M +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.2EH9LJehGa +++ cat /tmp/tmp.eZd8tzkM9M +++ rm /tmp/tmp.2EH9LJehGa /tmp/tmp.eZd8tzkM9M +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Yk31dBL2Py +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.eihENCckiB ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.Yk31dBL2Py ++++ cat /tmp/tmp.eihENCckiB ++++ rm /tmp/tmp.Yk31dBL2Py /tmp/tmp.eihENCckiB ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.2FmE0H33ON ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gFpbA7ia1U +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.2FmE0H33ON +++ cat /tmp/tmp.gFpbA7ia1U +++ rm /tmp/tmp.2FmE0H33ON /tmp/tmp.gFpbA7ia1U +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z9zVuDy6l9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.xckVYgsFHD ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.Z9zVuDy6l9 ++ cat /tmp/tmp.xckVYgsFHD ++ rm /tmp/tmp.Z9zVuDy6l9 /tmp/tmp.xckVYgsFHD ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.wwezathT5K ++ mktemp + local LAST_ERR=/tmp/tmp.fHTP7VcsEL + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.wwezathT5K + cat /tmp/tmp.fHTP7VcsEL + rm /tmp/tmp.wwezathT5K /tmp/tmp.fHTP7VcsEL + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WLurpKApwO ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fsLgIOS5z4 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.WLurpKApwO +++ cat /tmp/tmp.fsLgIOS5z4 +++ rm /tmp/tmp.WLurpKApwO /tmp/tmp.fsLgIOS5z4 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.nafnWwFk99 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BBovZXQM34 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.nafnWwFk99 +++ cat /tmp/tmp.BBovZXQM34 +++ rm /tmp/tmp.nafnWwFk99 /tmp/tmp.BBovZXQM34 +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.bCTQNZryEa ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Dvu1R8DLVd +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.bCTQNZryEa +++ cat /tmp/tmp.Dvu1R8DLVd +++ rm /tmp/tmp.bCTQNZryEa /tmp/tmp.Dvu1R8DLVd +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.evr9rV1ibQ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.hv1U0UPiAX ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.evr9rV1ibQ ++++ cat /tmp/tmp.hv1U0UPiAX ++++ rm /tmp/tmp.evr9rV1ibQ /tmp/tmp.hv1U0UPiAX ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hJxg8LfETX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.genlRXbrsa +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.hJxg8LfETX +++ cat /tmp/tmp.genlRXbrsa +++ rm /tmp/tmp.hJxg8LfETX /tmp/tmp.genlRXbrsa +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TX7Cj8xB8h +++ mktemp ++ local LAST_ERR=/tmp/tmp.QonVCyTPdk ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.TX7Cj8xB8h ++ cat /tmp/tmp.QonVCyTPdk ++ rm /tmp/tmp.TX7Cj8xB8h /tmp/tmp.QonVCyTPdk ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.NATZnR7f7N ++ mktemp + local LAST_ERR=/tmp/tmp.EZOMdxmdjk + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.NATZnR7f7N + cat /tmp/tmp.EZOMdxmdjk + rm /tmp/tmp.NATZnR7f7N /tmp/tmp.EZOMdxmdjk + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tS6myC2Ss5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.3mzGl4IFnj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tS6myC2Ss5 +++ cat /tmp/tmp.3mzGl4IFnj +++ rm /tmp/tmp.tS6myC2Ss5 /tmp/tmp.3mzGl4IFnj +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rmzez4xKeq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Ydu82arnAG +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rmzez4xKeq +++ cat /tmp/tmp.Ydu82arnAG +++ rm /tmp/tmp.rmzez4xKeq /tmp/tmp.Ydu82arnAG +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qNq2MY6WLm ++++ mktemp +++ local LAST_ERR=/tmp/tmp.UYey7NxE7G +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qNq2MY6WLm +++ cat /tmp/tmp.UYey7NxE7G +++ rm /tmp/tmp.qNq2MY6WLm /tmp/tmp.UYey7NxE7G +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.EpHviu2PL1 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.3Y9V1rpOy6 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.EpHviu2PL1 ++++ cat /tmp/tmp.3Y9V1rpOy6 ++++ rm /tmp/tmp.EpHviu2PL1 /tmp/tmp.3Y9V1rpOy6 ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.aoEehPVpPM ++++ mktemp +++ local LAST_ERR=/tmp/tmp.2dLxGGxHbO +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.aoEehPVpPM +++ cat /tmp/tmp.2dLxGGxHbO +++ rm /tmp/tmp.aoEehPVpPM /tmp/tmp.2dLxGGxHbO +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YXuMyOAB6H +++ mktemp ++ local LAST_ERR=/tmp/tmp.0labOMD5eW ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.YXuMyOAB6H ++ cat /tmp/tmp.0labOMD5eW ++ rm /tmp/tmp.YXuMyOAB6H /tmp/tmp.0labOMD5eW ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.FGTw9A87c7 ++ mktemp + local LAST_ERR=/tmp/tmp.AeW2xQxbIJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FGTw9A87c7 + cat /tmp/tmp.AeW2xQxbIJ + rm /tmp/tmp.FGTw9A87c7 /tmp/tmp.AeW2xQxbIJ + return 0 + sleep 10 + apply_version cluster1 9.9.9 12-latest 12 http://version-service:11000 3 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-latest + local db_version=12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=3 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.gTaHHpp61n ++ mktemp + local LAST_ERR=/tmp/tmp.feiMp5iHit + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gTaHHpp61n perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.feiMp5iHit + rm /tmp/tmp.gTaHHpp61n /tmp/tmp.feiMp5iHit + return 0 + sleep 10 + wait_smart_update cluster1 3 + local cluster=cluster1 + local target=3 + set +o xtrace ...................................................... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.WsMxp9niVx ++ mktemp + local LAST_ERR=/tmp/tmp.j4lP9IzCVt + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.WsMxp9niVx job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.j4lP9IzCVt + rm /tmp/tmp.WsMxp9niVx /tmp/tmp.j4lP9IzCVt + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.parameters.pg-cluster cluster1 ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' +++ mktemp + kubectl_bin apply -f - + yq w - spec.namespace version-service-7914 + yq w --style=single -- - spec.parameters.backrest-opts --type=full ++ local LAST_OUT=/tmp/tmp.Xd5TUCWeWW ++ mktemp +++ mktemp ++ local LAST_ERR=/tmp/tmp.OuNdiDvsth ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + yq w - metadata.name fulldata-cluster1 + local LAST_OUT=/tmp/tmp.YfNctIr8ww +++ seq 0 2 ++ mktemp + local LAST_ERR=/tmp/tmp.ndTImAKKZi + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.Xd5TUCWeWW ++ cat /tmp/tmp.OuNdiDvsth ++ rm /tmp/tmp.Xd5TUCWeWW /tmp/tmp.OuNdiDvsth ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-5f647d77b5-2b5rr + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.YfNctIr8ww pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.ndTImAKKZi + rm /tmp/tmp.YfNctIr8ww /tmp/tmp.ndTImAKKZi + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 12-latest 12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-latest + local db_version=12 + sort -s -u + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' ++ mktemp + local LAST_OUT=/tmp/tmp.tHNP3Paumu ++ mktemp + local LAST_ERR=/tmp/tmp.FClJJDPQj6 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.tHNP3Paumu + cat /tmp/tmp.FClJJDPQj6 + rm /tmp/tmp.tHNP3Paumu /tmp/tmp.FClJJDPQj6 + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-12-latest-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster update with explicitly specified image inside version service' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster update with explicitly specified image inside version service ----------------------------------------------------------------------------------- + PG_VER=13 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.yUhO5ZUTVN ++ mktemp + local LAST_ERR=/tmp/tmp.YXL0FzgIMq + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yUhO5ZUTVN deployment.apps/pg-client unchanged + cat /tmp/tmp.YXL0FzgIMq + rm /tmp/tmp.yUhO5ZUTVN /tmp/tmp.YXL0FzgIMq + return 0 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.GiGus99Fot ++ mktemp + local LAST_ERR=/tmp/tmp.XnHquurYos + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.GiGus99Fot secret/cluster1-users-secret created + cat /tmp/tmp.XnHquurYos + rm /tmp/tmp.GiGus99Fot /tmp/tmp.XnHquurYos + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ZPH96t2nfu ++ mktemp + local LAST_ERR=/tmp/tmp.G7pGTEYYdJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ZPH96t2nfu secret/cluster1-pmm-secret configured + cat /tmp/tmp.G7pGTEYYdJ + rm /tmp/tmp.ZPH96t2nfu /tmp/tmp.G7pGTEYYdJ + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.DvufLafxGT ++ mktemp + local LAST_ERR=/tmp/tmp.0p1ttO3JWS + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.DvufLafxGT deployment.apps/pg-client unchanged + cat /tmp/tmp.0p1ttO3JWS + rm /tmp/tmp.DvufLafxGT /tmp/tmp.0p1ttO3JWS + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + yq w - spec.clustername cluster1 + yq w - spec.database cluster1 + yq w - spec.name cluster1 + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + yq w - spec.keepData false + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + yq w - spec.user some-name + yq w - metadata.labels.deployment-name cluster1 + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + yq w - spec.keepBackups false + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.Z6ZfrH04VK ++ mktemp + local LAST_ERR=/tmp/tmp.q4MzuJsJQj + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Z6ZfrH04VK perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.q4MzuJsJQj + rm /tmp/tmp.Z6ZfrH04VK /tmp/tmp.q4MzuJsJQj + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1......... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DPUVeWkncx +++ mktemp ++ local LAST_ERR=/tmp/tmp.EwwnZwjnfV ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.DPUVeWkncx ++ cat /tmp/tmp.EwwnZwjnfV ++ rm /tmp/tmp.DPUVeWkncx /tmp/tmp.EwwnZwjnfV ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.U4LDEMFMmK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fbTnwhE0bS +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.U4LDEMFMmK +++ cat /tmp/tmp.fbTnwhE0bS +++ rm /tmp/tmp.U4LDEMFMmK /tmp/tmp.fbTnwhE0bS +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.C8FyPaXWar ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zkwznOhvFg +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.C8FyPaXWar +++ cat /tmp/tmp.zkwznOhvFg +++ rm /tmp/tmp.C8FyPaXWar /tmp/tmp.zkwznOhvFg +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.z4zT9jpWdS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vhdBpbcOxr +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.z4zT9jpWdS +++ cat /tmp/tmp.vhdBpbcOxr +++ rm /tmp/tmp.z4zT9jpWdS /tmp/tmp.vhdBpbcOxr +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.vOQuEdceSw ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wGhWju3R2Z +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.vOQuEdceSw +++ cat /tmp/tmp.wGhWju3R2Z +++ rm /tmp/tmp.vOQuEdceSw /tmp/tmp.wGhWju3R2Z +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.iGLyZzkyJc +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.lgRnbXk8hk ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.iGLyZzkyJc ++++ cat /tmp/tmp.lgRnbXk8hk ++++ rm /tmp/tmp.iGLyZzkyJc /tmp/tmp.lgRnbXk8hk ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dH6PTiagr5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uTL6DtRxs8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.dH6PTiagr5 +++ cat /tmp/tmp.uTL6DtRxs8 +++ rm /tmp/tmp.dH6PTiagr5 /tmp/tmp.uTL6DtRxs8 +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ch7Qp9nd2h +++ mktemp ++ local LAST_ERR=/tmp/tmp.e9pTDXVAI4 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.Ch7Qp9nd2h ++ cat /tmp/tmp.e9pTDXVAI4 ++ rm /tmp/tmp.Ch7Qp9nd2h /tmp/tmp.e9pTDXVAI4 ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.3sxhsmKYlo ++ mktemp + local LAST_ERR=/tmp/tmp.tMIi83dTdd + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.3sxhsmKYlo + cat /tmp/tmp.tMIi83dTdd + rm /tmp/tmp.3sxhsmKYlo /tmp/tmp.tMIi83dTdd + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mtSoYuT69q ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ow5IGhmzJW +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.mtSoYuT69q +++ cat /tmp/tmp.ow5IGhmzJW +++ rm /tmp/tmp.mtSoYuT69q /tmp/tmp.ow5IGhmzJW +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rlIhBq1auA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.m1gsOUrTPY +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rlIhBq1auA +++ cat /tmp/tmp.m1gsOUrTPY +++ rm /tmp/tmp.rlIhBq1auA /tmp/tmp.m1gsOUrTPY +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DUli5Cferh ++++ mktemp +++ local LAST_ERR=/tmp/tmp.N9EuPzL2Rs +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.DUli5Cferh +++ cat /tmp/tmp.N9EuPzL2Rs +++ rm /tmp/tmp.DUli5Cferh /tmp/tmp.N9EuPzL2Rs +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.0cSRGVWlG7 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.T0osmJPwuU ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.0cSRGVWlG7 ++++ cat /tmp/tmp.T0osmJPwuU ++++ rm /tmp/tmp.0cSRGVWlG7 /tmp/tmp.T0osmJPwuU ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rop4HH2Ydx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bv3TWAekzw +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rop4HH2Ydx +++ cat /tmp/tmp.bv3TWAekzw +++ rm /tmp/tmp.rop4HH2Ydx /tmp/tmp.bv3TWAekzw +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WnDr7Z4beD +++ mktemp ++ local LAST_ERR=/tmp/tmp.jXaX0EaBIF ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.WnDr7Z4beD ++ cat /tmp/tmp.jXaX0EaBIF ++ rm /tmp/tmp.WnDr7Z4beD /tmp/tmp.jXaX0EaBIF ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.JE71b3VBKM ++ mktemp + local LAST_ERR=/tmp/tmp.GtgipHhxVC + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.JE71b3VBKM + cat /tmp/tmp.GtgipHhxVC + rm /tmp/tmp.JE71b3VBKM /tmp/tmp.GtgipHhxVC + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.fPpKu9gFDS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.J0bXKDCadC +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.fPpKu9gFDS +++ cat /tmp/tmp.J0bXKDCadC +++ rm /tmp/tmp.fPpKu9gFDS /tmp/tmp.J0bXKDCadC +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.eQoLKuQpTq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.JyGVH2fgoH +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.eQoLKuQpTq +++ cat /tmp/tmp.JyGVH2fgoH +++ rm /tmp/tmp.eQoLKuQpTq /tmp/tmp.JyGVH2fgoH +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.cXMhbb4ix2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vtbX1PcgsH +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.cXMhbb4ix2 +++ cat /tmp/tmp.vtbX1PcgsH +++ rm /tmp/tmp.cXMhbb4ix2 /tmp/tmp.vtbX1PcgsH +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.iBVjhwMgGI +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.2393OMeC0o ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.iBVjhwMgGI ++++ cat /tmp/tmp.2393OMeC0o ++++ rm /tmp/tmp.iBVjhwMgGI /tmp/tmp.2393OMeC0o ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.zEVsCtTnw7 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TJ9jMggGz2 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.zEVsCtTnw7 +++ cat /tmp/tmp.TJ9jMggGz2 +++ rm /tmp/tmp.zEVsCtTnw7 /tmp/tmp.TJ9jMggGz2 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fFazrDGbKO +++ mktemp ++ local LAST_ERR=/tmp/tmp.TFdqd0idrj ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.fFazrDGbKO ++ cat /tmp/tmp.TFdqd0idrj ++ rm /tmp/tmp.fFazrDGbKO /tmp/tmp.TFdqd0idrj ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.Tej8e9LTTK ++ mktemp + local LAST_ERR=/tmp/tmp.aKtyNggjUT + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Tej8e9LTTK + cat /tmp/tmp.aKtyNggjUT + rm /tmp/tmp.Tej8e9LTTK /tmp/tmp.aKtyNggjUT + return 0 + sleep 10 + apply_version cluster1 9.9.9 13.14 13 http://version-service:11000 4 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=13.14 + local db_version=13 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=4 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"13.14"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.nrvqwnLW0r ++ mktemp + local LAST_ERR=/tmp/tmp.UMhzhwWgaN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"13.14"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nrvqwnLW0r perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.UMhzhwWgaN + rm /tmp/tmp.nrvqwnLW0r /tmp/tmp.UMhzhwWgaN + return 0 + sleep 10 + wait_smart_update cluster1 4 + local cluster=cluster1 + local target=4 + set +o xtrace .............................................. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.e2HHAOQAeR ++ mktemp + local LAST_ERR=/tmp/tmp.zkTSeOOSwr + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.e2HHAOQAeR job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.zkTSeOOSwr + rm /tmp/tmp.e2HHAOQAeR /tmp/tmp.zkTSeOOSwr + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + kubectl_bin apply -f - ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ mktemp + yq w - spec.namespace version-service-7914 +++ mktemp + yq w - spec.parameters.pg-cluster cluster1 + local LAST_OUT=/tmp/tmp.dYe3Dp76VG ++ local LAST_OUT=/tmp/tmp.KO2iScytim ++ mktemp +++ mktemp + local LAST_ERR=/tmp/tmp.mU0sdUQ8xO + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ local LAST_ERR=/tmp/tmp.mmzwn05QZy + for i in '$(seq 0 2)' ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.KO2iScytim ++ cat /tmp/tmp.mmzwn05QZy ++ rm /tmp/tmp.KO2iScytim /tmp/tmp.mmzwn05QZy ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-57f668d478-258cw + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.dYe3Dp76VG pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.mU0sdUQ8xO + rm /tmp/tmp.dYe3Dp76VG /tmp/tmp.mU0sdUQ8xO + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 13.14 13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=13.14 + local db_version=13 + sort -s -u + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' ++ mktemp + local LAST_OUT=/tmp/tmp.gU21fDn6H8 ++ mktemp + local LAST_ERR=/tmp/tmp.5ul94WWWtM + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gU21fDn6H8 + cat /tmp/tmp.5ul94WWWtM + rm /tmp/tmp.gU21fDn6H8 /tmp/tmp.5ul94WWWtM + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-13.14-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster with version service offline' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster with version service offline ----------------------------------------------------------------------------------- + PG_VER=13 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.uZ4nCDVTLu ++ mktemp + local LAST_ERR=/tmp/tmp.TKGmVNChdd + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.uZ4nCDVTLu secret/cluster1-users-secret created + cat /tmp/tmp.TKGmVNChdd + rm /tmp/tmp.uZ4nCDVTLu /tmp/tmp.TKGmVNChdd + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.kBh9gq1JAK ++ mktemp + local LAST_ERR=/tmp/tmp.fxjEofjof5 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.kBh9gq1JAK secret/cluster1-pmm-secret configured + cat /tmp/tmp.fxjEofjof5 + rm /tmp/tmp.kBh9gq1JAK /tmp/tmp.fxjEofjof5 + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9lBIz8c9zL ++ mktemp + local LAST_ERR=/tmp/tmp.Tiy8cd20bU + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9lBIz8c9zL deployment.apps/pg-client unchanged + cat /tmp/tmp.Tiy8cd20bU + rm /tmp/tmp.9lBIz8c9zL /tmp/tmp.Tiy8cd20bU + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.name cluster1 + yq w - spec.clustername cluster1 + yq w - spec.database cluster1 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer + yq w - spec.secretsName cluster1-users-secret + yq w - spec.user some-name + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger + yq w - spec.keepBackups false + yq w - spec.keepData false + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.qMHnzGZeCP ++ mktemp + local LAST_ERR=/tmp/tmp.KEjFtcFDL1 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.qMHnzGZeCP perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.KEjFtcFDL1 + rm /tmp/tmp.qMHnzGZeCP /tmp/tmp.KEjFtcFDL1 + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1....... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uujN4tF0Ul +++ mktemp ++ local LAST_ERR=/tmp/tmp.bSamDT0hzx ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.uujN4tF0Ul ++ cat /tmp/tmp.bSamDT0hzx ++ rm /tmp/tmp.uujN4tF0Ul /tmp/tmp.bSamDT0hzx ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.txZCJfjcfx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bqNG0e7PJ0 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.txZCJfjcfx +++ cat /tmp/tmp.bqNG0e7PJ0 +++ rm /tmp/tmp.txZCJfjcfx /tmp/tmp.bqNG0e7PJ0 +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.t16p3FrcoY ++++ mktemp +++ local LAST_ERR=/tmp/tmp.udzmf4lwdn +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.t16p3FrcoY +++ cat /tmp/tmp.udzmf4lwdn +++ rm /tmp/tmp.t16p3FrcoY /tmp/tmp.udzmf4lwdn +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.QE5S94x8eN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yHnxTN3AnT +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.QE5S94x8eN +++ cat /tmp/tmp.yHnxTN3AnT +++ rm /tmp/tmp.QE5S94x8eN /tmp/tmp.yHnxTN3AnT +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.xMykfwrSSy ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Zu2rIbScfY +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.xMykfwrSSy +++ cat /tmp/tmp.Zu2rIbScfY +++ rm /tmp/tmp.xMykfwrSSy /tmp/tmp.Zu2rIbScfY +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.pW8AHDI32T +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Pkg46MoV16 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.pW8AHDI32T ++++ cat /tmp/tmp.Pkg46MoV16 ++++ rm /tmp/tmp.pW8AHDI32T /tmp/tmp.Pkg46MoV16 ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DLDYwHDerk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.UTIjXJ3GH7 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.DLDYwHDerk +++ cat /tmp/tmp.UTIjXJ3GH7 +++ rm /tmp/tmp.DLDYwHDerk /tmp/tmp.UTIjXJ3GH7 +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TIwRPb883T +++ mktemp ++ local LAST_ERR=/tmp/tmp.MtMndr40IU ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.TIwRPb883T ++ cat /tmp/tmp.MtMndr40IU ++ rm /tmp/tmp.TIwRPb883T /tmp/tmp.MtMndr40IU ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.hhpcZ4nqXZ ++ mktemp + local LAST_ERR=/tmp/tmp.vLJpDx3j8J + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.hhpcZ4nqXZ + cat /tmp/tmp.vLJpDx3j8J + rm /tmp/tmp.hhpcZ4nqXZ /tmp/tmp.vLJpDx3j8J + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.UkErgDxPQq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.l4rU7ZCn0e +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.UkErgDxPQq +++ cat /tmp/tmp.l4rU7ZCn0e +++ rm /tmp/tmp.UkErgDxPQq /tmp/tmp.l4rU7ZCn0e +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.yy8qWxZ3IP ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NujFaoH4Fh +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.yy8qWxZ3IP +++ cat /tmp/tmp.NujFaoH4Fh +++ rm /tmp/tmp.yy8qWxZ3IP /tmp/tmp.NujFaoH4Fh +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.66qsXm875V ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ng4vD6l4G4 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.66qsXm875V +++ cat /tmp/tmp.ng4vD6l4G4 +++ rm /tmp/tmp.66qsXm875V /tmp/tmp.ng4vD6l4G4 +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.qqVbKle2Ts +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.sx9COtr76q ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.qqVbKle2Ts ++++ cat /tmp/tmp.sx9COtr76q ++++ rm /tmp/tmp.qqVbKle2Ts /tmp/tmp.sx9COtr76q ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.k9ZDMa2sNT ++++ mktemp +++ local LAST_ERR=/tmp/tmp.lAL4NmQbQT +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.k9ZDMa2sNT +++ cat /tmp/tmp.lAL4NmQbQT +++ rm /tmp/tmp.k9ZDMa2sNT /tmp/tmp.lAL4NmQbQT +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gJy1EIxgif +++ mktemp ++ local LAST_ERR=/tmp/tmp.NkOph4m5Ox ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.gJy1EIxgif ++ cat /tmp/tmp.NkOph4m5Ox ++ rm /tmp/tmp.gJy1EIxgif /tmp/tmp.NkOph4m5Ox ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.0m5GRXBb0b ++ mktemp + local LAST_ERR=/tmp/tmp.j77EiuW4pl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.0m5GRXBb0b + cat /tmp/tmp.j77EiuW4pl + rm /tmp/tmp.0m5GRXBb0b /tmp/tmp.j77EiuW4pl + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.XXtnrrh9td ++++ mktemp +++ local LAST_ERR=/tmp/tmp.xiApHxiD7q +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.XXtnrrh9td +++ cat /tmp/tmp.xiApHxiD7q +++ rm /tmp/tmp.XXtnrrh9td /tmp/tmp.xiApHxiD7q +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.4y0ehYD3Sv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.icv4veGYbT +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.4y0ehYD3Sv +++ cat /tmp/tmp.icv4veGYbT +++ rm /tmp/tmp.4y0ehYD3Sv /tmp/tmp.icv4veGYbT +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.oun9yUa1Zo ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZsITuXWrnU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.oun9yUa1Zo +++ cat /tmp/tmp.ZsITuXWrnU +++ rm /tmp/tmp.oun9yUa1Zo /tmp/tmp.ZsITuXWrnU +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.DP7Iveyvt1 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.BtXD3nuQHx ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.DP7Iveyvt1 ++++ cat /tmp/tmp.BtXD3nuQHx ++++ rm /tmp/tmp.DP7Iveyvt1 /tmp/tmp.BtXD3nuQHx ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.e6vZXilIVJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Zf1nq66oTx +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.e6vZXilIVJ +++ cat /tmp/tmp.Zf1nq66oTx +++ rm /tmp/tmp.e6vZXilIVJ /tmp/tmp.Zf1nq66oTx +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h8KMWZlhKj +++ mktemp ++ local LAST_ERR=/tmp/tmp.xUjlozW2KS ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.h8KMWZlhKj ++ cat /tmp/tmp.xUjlozW2KS ++ rm /tmp/tmp.h8KMWZlhKj /tmp/tmp.xUjlozW2KS ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.0pIXrCuUPc ++ mktemp + local LAST_ERR=/tmp/tmp.pUIj4fwSeJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.0pIXrCuUPc + cat /tmp/tmp.pUIj4fwSeJ + rm /tmp/tmp.0pIXrCuUPc /tmp/tmp.pUIj4fwSeJ + return 0 + sleep 10 + apply_version cluster1 9.9.9 latest 1.6.0-13 http://not-a-version-service/ 4 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=1.6.0-13 + local vs_endpoint=http://not-a-version-service/ + local update_finished_log_entries=4 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://not-a-version-service/","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.NQYTcTFBTD ++ mktemp + local LAST_ERR=/tmp/tmp.hEeT0tLcx9 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://not-a-version-service/","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.NQYTcTFBTD perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.hEeT0tLcx9 + rm /tmp/tmp.NQYTcTFBTD /tmp/tmp.hEeT0tLcx9 + return 0 + sleep 10 + wait_smart_update cluster1 4 + local cluster=cluster1 + local target=4 + set +o xtrace + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1................. + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.gZM1dhb7dg ++ mktemp + local LAST_ERR=/tmp/tmp.haRu0hefRQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gZM1dhb7dg job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.haRu0hefRQ + rm /tmp/tmp.gZM1dhb7dg /tmp/tmp.haRu0hefRQ + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.parameters.pg-cluster cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 + kubectl_bin apply -f - + yq w --style=single -- - spec.parameters.backrest-opts --type=full ++ mktemp + yq w - spec.namespace version-service-7914 + local LAST_OUT=/tmp/tmp.yaocsfuSFr ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ mktemp +++ mktemp + local LAST_ERR=/tmp/tmp.5cZ9Wln0gY + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ local LAST_OUT=/tmp/tmp.XCOzUDrWE8 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - +++ mktemp ++ local LAST_ERR=/tmp/tmp.qIKF70kM3E ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.XCOzUDrWE8 ++ cat /tmp/tmp.qIKF70kM3E ++ rm /tmp/tmp.XCOzUDrWE8 /tmp/tmp.qIKF70kM3E ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-9f8cd5549-8sb26 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yaocsfuSFr pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.5cZ9Wln0gY + rm /tmp/tmp.yaocsfuSFr /tmp/tmp.5cZ9Wln0gY + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 latest 1.6.0-13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=1.6.0-13 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.3PwK10Wwcb ++ mktemp + local LAST_ERR=/tmp/tmp.INwcUkzSlW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.3PwK10Wwcb + cat /tmp/tmp.INwcUkzSlW + rm /tmp/tmp.3PwK10Wwcb /tmp/tmp.INwcUkzSlW + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-latest-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/1.6.0-13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Updating PG cluster with version service available but disabled' + set +o xtrace ----------------------------------------------------------------------------------- Updating PG cluster with version service available but disabled ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-7914 apply -f - + yq w - metadata.name cluster1-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.z3YwLHDc3J ++ mktemp + local LAST_ERR=/tmp/tmp.mkRHbx0wI3 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.z3YwLHDc3J secret/cluster1-users-secret created + cat /tmp/tmp.mkRHbx0wI3 + rm /tmp/tmp.z3YwLHDc3J /tmp/tmp.mkRHbx0wI3 + return 0 + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.GdATO8FhSm ++ mktemp + local LAST_ERR=/tmp/tmp.NQXinFygR6 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.GdATO8FhSm secret/cluster1-pmm-secret configured + cat /tmp/tmp.NQXinFygR6 + rm /tmp/tmp.GdATO8FhSm /tmp/tmp.NQXinFygR6 + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.EhPXus9S4R ++ mktemp + local LAST_ERR=/tmp/tmp.lI3k8Z1N6Y + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.EhPXus9S4R deployment.apps/pg-client unchanged + cat /tmp/tmp.lI3k8Z1N6Y + rm /tmp/tmp.EhPXus9S4R /tmp/tmp.lI3k8Z1N6Y + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.name cluster1 + yq w - spec.user some-name + yq w - spec.clustername cluster1 + yq w - spec.database cluster1 + yq w - spec.name cluster1 + yq w - spec.secretsName cluster1-users-secret + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + yq w - spec.keepData false + yq w - spec.keepBackups false + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.qWC24S8pu0 ++ mktemp + local LAST_ERR=/tmp/tmp.2DIlaJD72E + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.qWC24S8pu0 perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.2DIlaJD72E + rm /tmp/tmp.qWC24S8pu0 /tmp/tmp.2DIlaJD72E + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1......... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3wsN6F4v2k +++ mktemp ++ local LAST_ERR=/tmp/tmp.QACR1IFUoI ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.3wsN6F4v2k ++ cat /tmp/tmp.QACR1IFUoI ++ rm /tmp/tmp.3wsN6F4v2k /tmp/tmp.QACR1IFUoI ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.z6MtVEV95E ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yhEHt7rV73 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.z6MtVEV95E +++ cat /tmp/tmp.yhEHt7rV73 +++ rm /tmp/tmp.z6MtVEV95E /tmp/tmp.yhEHt7rV73 +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.XVpBoYx2k9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.SKswZFGRuj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.XVpBoYx2k9 +++ cat /tmp/tmp.SKswZFGRuj +++ rm /tmp/tmp.XVpBoYx2k9 /tmp/tmp.SKswZFGRuj +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.XNgCnGl5Dz ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VRI0ilIQwA +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.XNgCnGl5Dz +++ cat /tmp/tmp.VRI0ilIQwA +++ rm /tmp/tmp.XNgCnGl5Dz /tmp/tmp.VRI0ilIQwA +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.gVhNRMvpFm ++++ mktemp +++ local LAST_ERR=/tmp/tmp.S5oiyQblOY +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.gVhNRMvpFm +++ cat /tmp/tmp.S5oiyQblOY +++ rm /tmp/tmp.gVhNRMvpFm /tmp/tmp.S5oiyQblOY +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.CQ0LpRtsLx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yiREaSjYWU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.CQ0LpRtsLx +++ cat /tmp/tmp.yiREaSjYWU +++ rm /tmp/tmp.CQ0LpRtsLx /tmp/tmp.yiREaSjYWU +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.DZyIzIGtGl +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.UxbIpkSrUZ ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.DZyIzIGtGl ++++ cat /tmp/tmp.UxbIpkSrUZ ++++ rm /tmp/tmp.DZyIzIGtGl /tmp/tmp.UxbIpkSrUZ ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.glKnQeSAXf ++++ mktemp +++ local LAST_ERR=/tmp/tmp.APb6Vlvsl3 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.glKnQeSAXf +++ cat /tmp/tmp.APb6Vlvsl3 +++ rm /tmp/tmp.glKnQeSAXf /tmp/tmp.APb6Vlvsl3 +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.doYZw3I7aW +++ mktemp ++ local LAST_ERR=/tmp/tmp.LKsWe0Ifjt ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.doYZw3I7aW ++ cat /tmp/tmp.LKsWe0Ifjt ++ rm /tmp/tmp.doYZw3I7aW /tmp/tmp.LKsWe0Ifjt ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.iO8ORjHtyO ++ mktemp + local LAST_ERR=/tmp/tmp.ZkLtQcdOaW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.iO8ORjHtyO + cat /tmp/tmp.ZkLtQcdOaW + rm /tmp/tmp.iO8ORjHtyO /tmp/tmp.ZkLtQcdOaW + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lpOUT1QvuE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.x0YrZB42ax +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.lpOUT1QvuE +++ cat /tmp/tmp.x0YrZB42ax +++ rm /tmp/tmp.lpOUT1QvuE /tmp/tmp.x0YrZB42ax +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VUWpt5K9ew ++++ mktemp +++ local LAST_ERR=/tmp/tmp.JMMD2JSdeR +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.VUWpt5K9ew +++ cat /tmp/tmp.JMMD2JSdeR +++ rm /tmp/tmp.VUWpt5K9ew /tmp/tmp.JMMD2JSdeR +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wLKPh6Q9MN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.P7EumPy9XO +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.wLKPh6Q9MN +++ cat /tmp/tmp.P7EumPy9XO +++ rm /tmp/tmp.wLKPh6Q9MN /tmp/tmp.P7EumPy9XO +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.4m76fxe6A6 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.PFPFYzNre7 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.4m76fxe6A6 ++++ cat /tmp/tmp.PFPFYzNre7 ++++ rm /tmp/tmp.4m76fxe6A6 /tmp/tmp.PFPFYzNre7 ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WIdGyGcbZM ++++ mktemp +++ local LAST_ERR=/tmp/tmp.emKuJi9Egp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.WIdGyGcbZM +++ cat /tmp/tmp.emKuJi9Egp +++ rm /tmp/tmp.WIdGyGcbZM /tmp/tmp.emKuJi9Egp +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QPll3gVL44 +++ mktemp ++ local LAST_ERR=/tmp/tmp.sfEjgYAOfn ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.QPll3gVL44 ++ cat /tmp/tmp.sfEjgYAOfn ++ rm /tmp/tmp.QPll3gVL44 /tmp/tmp.sfEjgYAOfn ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.yIUkBSkQQk ++ mktemp + local LAST_ERR=/tmp/tmp.ZvbHgfMxDa + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yIUkBSkQQk + cat /tmp/tmp.ZvbHgfMxDa + rm /tmp/tmp.yIUkBSkQQk /tmp/tmp.ZvbHgfMxDa + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.by04Eb1ABt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.RGJQny6k9F +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.by04Eb1ABt +++ cat /tmp/tmp.RGJQny6k9F +++ rm /tmp/tmp.by04Eb1ABt /tmp/tmp.RGJQny6k9F +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MYh4ntDTio ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0StgZaxVdD +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MYh4ntDTio +++ cat /tmp/tmp.0StgZaxVdD +++ rm /tmp/tmp.MYh4ntDTio /tmp/tmp.0StgZaxVdD +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.iVMJDG2dJ0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.FWBAdcm3nU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.iVMJDG2dJ0 +++ cat /tmp/tmp.FWBAdcm3nU +++ rm /tmp/tmp.iVMJDG2dJ0 /tmp/tmp.FWBAdcm3nU +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.rFKdYbLt7P +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.O1t0J66Ex5 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.rFKdYbLt7P ++++ cat /tmp/tmp.O1t0J66Ex5 ++++ rm /tmp/tmp.rFKdYbLt7P /tmp/tmp.O1t0J66Ex5 ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Wa7oUzRmrt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Bm3oWOTGTS +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Wa7oUzRmrt +++ cat /tmp/tmp.Bm3oWOTGTS +++ rm /tmp/tmp.Wa7oUzRmrt /tmp/tmp.Bm3oWOTGTS +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dDkevBT8o6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.oJVuMbwHDQ ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.dDkevBT8o6 ++ cat /tmp/tmp.oJVuMbwHDQ ++ rm /tmp/tmp.dDkevBT8o6 /tmp/tmp.oJVuMbwHDQ ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.T3boxwOSxf ++ mktemp + local LAST_ERR=/tmp/tmp.xEXG57rN8h + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.T3boxwOSxf + cat /tmp/tmp.xEXG57rN8h + rm /tmp/tmp.T3boxwOSxf /tmp/tmp.xEXG57rN8h + return 0 + sleep 10 + apply_version cluster1 9.9.9 disabled 1.6.0-12 http://version-service:11000 4 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=disabled + local db_version=1.6.0-12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=4 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"disabled"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.FYASBWhD8U ++ mktemp + local LAST_ERR=/tmp/tmp.G5OMdnaqnj + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"disabled"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FYASBWhD8U perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.G5OMdnaqnj + rm /tmp/tmp.FYASBWhD8U /tmp/tmp.G5OMdnaqnj + return 0 + sleep 10 + wait_smart_update cluster1 4 + local cluster=cluster1 + local target=4 + set +o xtrace + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1....... + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.76EreQQvA5 ++ mktemp + local LAST_ERR=/tmp/tmp.X1IeaDLBI4 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.76EreQQvA5 job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.X1IeaDLBI4 + rm /tmp/tmp.76EreQQvA5 /tmp/tmp.X1IeaDLBI4 + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - metadata.name fulldata-cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + yq w - spec.namespace version-service-7914 + yq w - spec.parameters.pg-cluster cluster1 + yq w - spec.name fulldata-cluster1 ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' + kubectl_bin apply -f - ++ mktemp +++ mktemp + local LAST_OUT=/tmp/tmp.tKN4qI6KW6 ++ mktemp ++ local LAST_OUT=/tmp/tmp.dHbXZ3P6tL +++ mktemp + local LAST_ERR=/tmp/tmp.RUtWCE9VMj + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f - ++ local LAST_ERR=/tmp/tmp.q6Sbu5y8zG ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.dHbXZ3P6tL ++ cat /tmp/tmp.q6Sbu5y8zG ++ rm /tmp/tmp.dHbXZ3P6tL /tmp/tmp.q6Sbu5y8zG ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-574b598fb6-tzkkt + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.tKN4qI6KW6 pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.RUtWCE9VMj + rm /tmp/tmp.tKN4qI6KW6 /tmp/tmp.RUtWCE9VMj + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 disabled 1.6.0-12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=disabled + local db_version=1.6.0-12 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.FgwB8ceKfT ++ mktemp + local LAST_ERR=/tmp/tmp.c7PJ9MHoaY + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FgwB8ceKfT + cat /tmp/tmp.c7PJ9MHoaY + rm /tmp/tmp.FgwB8ceKfT /tmp/tmp.c7PJ9MHoaY + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-disabled-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/1.6.0-12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Start the cluster up with version service images' + set +o xtrace ----------------------------------------------------------------------------------- Start the cluster up with version service images ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + yq w -i /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.labels.pgo-version 9.9.9 + yq w -i /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml spec.userLabels.pgo-version 9.9.9 + yq w -i /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml spec.upgradeOptions.versionServiceEndpoint http://version-service:11000 + yq w -i --style=single /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml spec.upgradeOptions.apply recommended + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=version-service-7914 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-7914 + local cluster=cluster1 + local namespace=version-service-7914 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + yq w - metadata.name cluster1-users-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.Z8qLIS0Qa5 ++ mktemp + local LAST_ERR=/tmp/tmp.HBQ4XAknME + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Z8qLIS0Qa5 secret/cluster1-users-secret created + cat /tmp/tmp.HBQ4XAknME + rm /tmp/tmp.Z8qLIS0Qa5 /tmp/tmp.HBQ4XAknME + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-7914 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.zyDEEVx6ZI ++ mktemp + local LAST_ERR=/tmp/tmp.ySQL66eFDM + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.zyDEEVx6ZI secret/cluster1-pmm-secret configured + cat /tmp/tmp.ySQL66eFDM + rm /tmp/tmp.zyDEEVx6ZI /tmp/tmp.ySQL66eFDM + return 0 + kubectl_bin -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.gHjPJ4JVES ++ mktemp + local LAST_ERR=/tmp/tmp.d2UNXoplSJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gHjPJ4JVES deployment.apps/pg-client unchanged + cat /tmp/tmp.d2UNXoplSJ + rm /tmp/tmp.gHjPJ4JVES /tmp/tmp.d2UNXoplSJ + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/conf/1.6.0_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.clustername cluster1 + yq w - metadata.labels.name cluster1 + yq w - spec.database cluster1 + yq w - spec.name cluster1 + yq w - spec.backup.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo + yq w - spec.keepData false + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.keepBackups false + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer + yq w - spec.user some-name + yq w - metadata.labels.deployment-name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 ++ mktemp + local LAST_OUT=/tmp/tmp.oRr8ExxjVH ++ mktemp + local LAST_ERR=/tmp/tmp.FoKvLiIUeK + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/cr.yaml -n version-service-7914 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.oRr8ExxjVH perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.FoKvLiIUeK + rm /tmp/tmp.oRr8ExxjVH /tmp/tmp.FoKvLiIUeK + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-7914 + sleep 10 + set +o xtrace cluster1..................... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-7914 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nps94cK1nD +++ mktemp ++ local LAST_ERR=/tmp/tmp.g7Cm6hi8Fr ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.nps94cK1nD ++ cat /tmp/tmp.g7Cm6hi8Fr ++ rm /tmp/tmp.nps94cK1nD /tmp/tmp.g7Cm6hi8Fr ++ return 0 + status_query='kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.9kDN8V2Vtt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZEOiXqxN4P +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.9kDN8V2Vtt +++ cat /tmp/tmp.ZEOiXqxN4P +++ rm /tmp/tmp.9kDN8V2Vtt /tmp/tmp.ZEOiXqxN4P +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.xi9Oltr4Au ++++ mktemp +++ local LAST_ERR=/tmp/tmp.e0IGN9k2sp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.xi9Oltr4Au +++ cat /tmp/tmp.e0IGN9k2sp +++ rm /tmp/tmp.xi9Oltr4Au /tmp/tmp.e0IGN9k2sp +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.M0X6yxREZj ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TWjNdAPgcA +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.M0X6yxREZj +++ cat /tmp/tmp.TWjNdAPgcA +++ rm /tmp/tmp.M0X6yxREZj /tmp/tmp.TWjNdAPgcA +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.S8GfzdOMSD ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1ME4zu5ICh +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.S8GfzdOMSD +++ cat /tmp/tmp.1ME4zu5ICh +++ rm /tmp/tmp.S8GfzdOMSD /tmp/tmp.1ME4zu5ICh +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.HDvJ8EcDsn +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.0aIxUokXje ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.HDvJ8EcDsn ++++ cat /tmp/tmp.0aIxUokXje ++++ rm /tmp/tmp.HDvJ8EcDsn /tmp/tmp.0aIxUokXje ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.GrzxQwe9Vj ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Ge5yVC0YZB +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.GrzxQwe9Vj +++ cat /tmp/tmp.Ge5yVC0YZB +++ rm /tmp/tmp.GrzxQwe9Vj /tmp/tmp.Ge5yVC0YZB +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-7914 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p6Lj0OGiS2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.23Gs8KqZzB ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.p6Lj0OGiS2 ++ cat /tmp/tmp.23Gs8KqZzB ++ rm /tmp/tmp.p6Lj0OGiS2 /tmp/tmp.23Gs8KqZzB ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.TPbvHCsS1N ++ mktemp + local LAST_ERR=/tmp/tmp.AaueB9kOqA + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.TPbvHCsS1N + cat /tmp/tmp.AaueB9kOqA + rm /tmp/tmp.TPbvHCsS1N /tmp/tmp.AaueB9kOqA + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.d9nMU60pMX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.EazM7UbNLH +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.d9nMU60pMX +++ cat /tmp/tmp.EazM7UbNLH +++ rm /tmp/tmp.d9nMU60pMX /tmp/tmp.EazM7UbNLH +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.TVvyMHyVku ++++ mktemp +++ local LAST_ERR=/tmp/tmp.tNW8xuKyGW +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.TVvyMHyVku +++ cat /tmp/tmp.tNW8xuKyGW +++ rm /tmp/tmp.TVvyMHyVku /tmp/tmp.tNW8xuKyGW +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.muJppDZhQr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.OV0i8TF3PW +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.muJppDZhQr +++ cat /tmp/tmp.OV0i8TF3PW +++ rm /tmp/tmp.muJppDZhQr /tmp/tmp.OV0i8TF3PW +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.oJVOLSdGFG +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.i4F6OHmz0P ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.oJVOLSdGFG ++++ cat /tmp/tmp.i4F6OHmz0P ++++ rm /tmp/tmp.oJVOLSdGFG /tmp/tmp.i4F6OHmz0P ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.FiSSSJ0RMZ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zbXhRksjqJ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.FiSSSJ0RMZ +++ cat /tmp/tmp.zbXhRksjqJ +++ rm /tmp/tmp.FiSSSJ0RMZ /tmp/tmp.zbXhRksjqJ +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i39Az1Xog2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.IZOUISIKqT ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.i39Az1Xog2 ++ cat /tmp/tmp.IZOUISIKqT ++ rm /tmp/tmp.i39Az1Xog2 /tmp/tmp.IZOUISIKqT ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.qhTF20vrtc ++ mktemp + local LAST_ERR=/tmp/tmp.6XpY5uakkL + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.qhTF20vrtc + cat /tmp/tmp.6XpY5uakkL + rm /tmp/tmp.qhTF20vrtc /tmp/tmp.6XpY5uakkL + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-7914 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.yEq8c84Wsa ++++ mktemp +++ local LAST_ERR=/tmp/tmp.nVXXKpQ6WX +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.yEq8c84Wsa +++ cat /tmp/tmp.nVXXKpQ6WX +++ rm /tmp/tmp.yEq8c84Wsa /tmp/tmp.nVXXKpQ6WX +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lqf8eO7Ius ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fkxhal904e +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.lqf8eO7Ius +++ cat /tmp/tmp.fkxhal904e +++ rm /tmp/tmp.lqf8eO7Ius /tmp/tmp.fkxhal904e +++ return 0 +++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hppRRO2lMz ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Go6SSfC6lU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.hppRRO2lMz +++ cat /tmp/tmp.Go6SSfC6lU +++ rm /tmp/tmp.hppRRO2lMz /tmp/tmp.Go6SSfC6lU +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.hwoMndTxqN +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.gk2EORDyrF ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++++ kubectl -n version-service-7914 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.hwoMndTxqN ++++ cat /tmp/tmp.gk2EORDyrF ++++ rm /tmp/tmp.hwoMndTxqN /tmp/tmp.gk2EORDyrF ++++ return 0 +++ kubectl_bin -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Sdtk3ZRgPn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5qlItJH4TB +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service +++ kubectl -n version-service-7914 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Sdtk3ZRgPn +++ cat /tmp/tmp.5qlItJH4TB +++ rm /tmp/tmp.Sdtk3ZRgPn /tmp/tmp.5qlItJH4TB +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-7914 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-7914 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q5yplSWT8q +++ mktemp ++ local LAST_ERR=/tmp/tmp.f933dxiZzq ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl -n version-service-7914 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.q5yplSWT8q ++ cat /tmp/tmp.f933dxiZzq ++ rm /tmp/tmp.q5yplSWT8q /tmp/tmp.f933dxiZzq ++ return 0 + local client_container=pg-client-76df9df876-jwrzl + kubectl_bin exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.DEfSBFYnyG ++ mktemp + local LAST_ERR=/tmp/tmp.omGqyJNj1g + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl exec pg-client-76df9df876-jwrzl -n version-service-7914 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-7914.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.DEfSBFYnyG + cat /tmp/tmp.omGqyJNj1g + rm /tmp/tmp.DEfSBFYnyG /tmp/tmp.omGqyJNj1g + return 0 + sleep 10 + verify_applied_version cluster1 9.9.9 recommended 14 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=recommended + local db_version=14 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.9tMaQ2rPPJ ++ mktemp + local LAST_ERR=/tmp/tmp.Hn4z4jpDqw + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9tMaQ2rPPJ + cat /tmp/tmp.Hn4z4jpDqw + rm /tmp/tmp.9tMaQ2rPPJ /tmp/tmp.Hn4z4jpDqw + return 0 + diff -u /tmp/tmp.IoigtEzcpK/9.9.9-deployed-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/version-service/compare/14.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + destroy version-service-7914 + grep -v level=debug + grep -v level=info ++ get_operator_pod ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' + grep -v 'Getting tasks for pod' + grep -v 'Getting pods from source' + grep -v 'get backup status: Job.batch' + sort -u +++ mktemp + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + grep -v 'the object has been modified' + tee /tmp/tmp.IoigtEzcpK/operator.log ++ local LAST_OUT=/tmp/tmp.HC686PSIXV +++ mktemp ++ local LAST_ERR=/tmp/tmp.wuCxV1yDam ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.HC686PSIXV ++ cat /tmp/tmp.wuCxV1yDam ++ rm /tmp/tmp.HC686PSIXV /tmp/tmp.wuCxV1yDam ++ return 0 + kubectl_bin logs postgres-operator-7857c9f44-fbkqk -c operator ++ mktemp + local LAST_OUT=/tmp/tmp.JpkFbBI4OH ++ mktemp + local LAST_ERR=/tmp/tmp.gP2r69E26R + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl logs postgres-operator-7857c9f44-fbkqk -c operator + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.JpkFbBI4OH + cat /tmp/tmp.gP2r69E26R + rm /tmp/tmp.JpkFbBI4OH /tmp/tmp.gP2r69E26R + return 0 ] { } }, ] } }, { } }, ] { } }, { } ] } { } }, ] } ] { } }, }, { ] { } }, ] ] ], { } }, }, { }, { },{ }] }], ] { } }, , ,{ ], { } }, }], { } }, }], } ], { } }, } ] ], } } ], }, }, { } }, { } }, }, 2024/05/24 18:49:53 INF 1 (localhost:4150) connecting to nsqd 2024/05/24 18:50:43 INF 2 (localhost:4150) connecting to nsqd 2024/05/24 18:50:44 INF 3 (localhost:4150) connecting to nsqd 2024/05/24 18:50:50 INF 4 (localhost:4150) connecting to nsqd 2024/05/24 18:51:18 INF 5 (localhost:4150) connecting to nsqd 2024/05/24 18:51:18 INF 6 (localhost:4150) connecting to nsqd 2024/05/24 18:51:19 INF 7 (localhost:4150) connecting to nsqd 2024/05/24 18:59:08 INF 8 (localhost:4150) connecting to nsqd 2024/05/24 19:00:45 INF 9 (localhost:4150) connecting to nsqd 2024/05/24 19:01:00 INF 10 (localhost:4150) connecting to nsqd 2024/05/24 19:01:03 INF 11 (localhost:4150) connecting to nsqd 2024/05/24 19:01:22 INF 12 (localhost:4150) connecting to nsqd 2024/05/24 19:01:27 INF 13 (localhost:4150) connecting to nsqd 2024/05/24 19:01:44 INF 14 (localhost:4150) connecting to nsqd 2024/05/24 19:02:34 INF 15 (localhost:4150) connecting to nsqd 2024/05/24 19:02:35 INF 16 (localhost:4150) connecting to nsqd 2024/05/24 19:02:42 INF 17 (localhost:4150) connecting to nsqd 2024/05/24 19:03:08 INF 18 (localhost:4150) connecting to nsqd 2024/05/24 19:03:09 INF 19 (localhost:4150) connecting to nsqd 2024/05/24 19:03:09 INF 20 (localhost:4150) connecting to nsqd 2024/05/24 19:07:21 INF 21 (localhost:4150) connecting to nsqd 2024/05/24 19:08:53 INF 22 (localhost:4150) connecting to nsqd 2024/05/24 19:09:08 INF 23 (localhost:4150) connecting to nsqd 2024/05/24 19:09:10 INF 24 (localhost:4150) connecting to nsqd 2024/05/24 19:09:34 INF 25 (localhost:4150) connecting to nsqd 2024/05/24 19:09:39 INF 26 (localhost:4150) connecting to nsqd 2024/05/24 19:09:49 INF 27 (localhost:4150) connecting to nsqd 2024/05/24 19:10:38 INF 28 (localhost:4150) connecting to nsqd 2024/05/24 19:10:38 INF 29 (localhost:4150) connecting to nsqd 2024/05/24 19:10:45 INF 30 (localhost:4150) connecting to nsqd 2024/05/24 19:11:13 INF 31 (localhost:4150) connecting to nsqd 2024/05/24 19:11:13 INF 32 (localhost:4150) connecting to nsqd 2024/05/24 19:11:14 INF 33 (localhost:4150) connecting to nsqd 2024/05/24 19:17:03 INF 34 (localhost:4150) connecting to nsqd 2024/05/24 19:17:18 INF 35 (localhost:4150) connecting to nsqd 2024/05/24 19:17:21 INF 36 (localhost:4150) connecting to nsqd 2024/05/24 19:17:41 INF 37 (localhost:4150) connecting to nsqd 2024/05/24 19:17:46 INF 38 (localhost:4150) connecting to nsqd 2024/05/24 19:17:59 INF 39 (localhost:4150) connecting to nsqd 2024/05/24 19:18:48 INF 40 (localhost:4150) connecting to nsqd 2024/05/24 19:18:49 INF 41 (localhost:4150) connecting to nsqd 2024/05/24 19:18:55 INF 42 (localhost:4150) connecting to nsqd 2024/05/24 19:19:22 INF 43 (localhost:4150) connecting to nsqd 2024/05/24 19:19:22 INF 44 (localhost:4150) connecting to nsqd 2024/05/24 19:19:22 INF 45 (localhost:4150) connecting to nsqd 2024/05/24 19:23:06 INF 46 (localhost:4150) connecting to nsqd 2024/05/24 19:24:42 INF 47 (localhost:4150) connecting to nsqd 2024/05/24 19:24:58 INF 48 (localhost:4150) connecting to nsqd 2024/05/24 19:25:00 INF 49 (localhost:4150) connecting to nsqd 2024/05/24 19:25:19 INF 50 (localhost:4150) connecting to nsqd 2024/05/24 19:25:24 INF 51 (localhost:4150) connecting to nsqd 2024/05/24 19:25:41 INF 52 (localhost:4150) connecting to nsqd 2024/05/24 19:26:27 INF 53 (localhost:4150) connecting to nsqd 2024/05/24 19:26:29 INF 54 (localhost:4150) connecting to nsqd 2024/05/24 19:26:35 INF 55 (localhost:4150) connecting to nsqd 2024/05/24 19:27:02 INF 56 (localhost:4150) connecting to nsqd 2024/05/24 19:27:03 INF 57 (localhost:4150) connecting to nsqd 2024/05/24 19:27:03 INF 58 (localhost:4150) connecting to nsqd 2024/05/24 19:31:14 INF 59 (localhost:4150) connecting to nsqd 2024/05/24 19:32:48 INF 60 (localhost:4150) connecting to nsqd 2024/05/24 19:33:14 INF 61 (localhost:4150) connecting to nsqd 2024/05/24 19:33:17 INF 62 (localhost:4150) connecting to nsqd 2024/05/24 19:33:35 INF 63 (localhost:4150) connecting to nsqd 2024/05/24 19:33:40 INF 64 (localhost:4150) connecting to nsqd 2024/05/24 19:33:59 INF 65 (localhost:4150) connecting to nsqd 2024/05/24 19:34:48 INF 66 (localhost:4150) connecting to nsqd 2024/05/24 19:34:49 INF 67 (localhost:4150) connecting to nsqd 2024/05/24 19:34:55 INF 68 (localhost:4150) connecting to nsqd 2024/05/24 19:35:21 INF 69 (localhost:4150) connecting to nsqd 2024/05/24 19:35:21 INF 70 (localhost:4150) connecting to nsqd 2024/05/24 19:35:22 INF 71 (localhost:4150) connecting to nsqd 2024/05/24 19:42:19 INF 72 (localhost:4150) connecting to nsqd 2024/05/24 19:43:56 INF 73 (localhost:4150) connecting to nsqd 2024/05/24 19:44:11 INF 74 (localhost:4150) connecting to nsqd 2024/05/24 19:44:14 INF 75 (localhost:4150) connecting to nsqd 2024/05/24 19:44:31 INF 76 (localhost:4150) connecting to nsqd 2024/05/24 19:44:36 INF 77 (localhost:4150) connecting to nsqd 2024/05/24 19:44:54 INF 78 (localhost:4150) connecting to nsqd 2024/05/24 19:45:41 INF 79 (localhost:4150) connecting to nsqd 2024/05/24 19:45:42 INF 80 (localhost:4150) connecting to nsqd 2024/05/24 19:45:49 INF 81 (localhost:4150) connecting to nsqd 2024/05/24 19:46:16 INF 82 (localhost:4150) connecting to nsqd 2024/05/24 19:46:16 INF 83 (localhost:4150) connecting to nsqd 2024/05/24 19:46:16 INF 84 (localhost:4150) connecting to nsqd 2024/05/24 19:49:43 INF 85 (localhost:4150) connecting to nsqd 2024/05/24 19:49:57 INF 86 (localhost:4150) connecting to nsqd 2024/05/24 19:50:00 INF 87 (localhost:4150) connecting to nsqd 2024/05/24 19:50:23 INF 88 (localhost:4150) connecting to nsqd 2024/05/24 19:50:28 INF 89 (localhost:4150) connecting to nsqd 2024/05/24 19:50:39 INF 90 (localhost:4150) connecting to nsqd 2024/05/24 19:51:27 INF 91 (localhost:4150) connecting to nsqd 2024/05/24 19:51:29 INF 92 (localhost:4150) connecting to nsqd 2024/05/24 19:51:35 INF 93 (localhost:4150) connecting to nsqd 2024/05/24 19:52:01 INF 94 (localhost:4150) connecting to nsqd 2024/05/24 19:52:01 INF 95 (localhost:4150) connecting to nsqd 2024/05/24 19:52:01 INF 96 (localhost:4150) connecting to nsqd 2024/05/24 19:55:09 INF 97 (localhost:4150) connecting to nsqd 2024/05/24 19:55:34 INF 98 (localhost:4150) connecting to nsqd 2024/05/24 19:55:37 INF 99 (localhost:4150) connecting to nsqd 2024/05/24 19:55:56 INF 100 (localhost:4150) connecting to nsqd 2024/05/24 19:56:01 INF 101 (localhost:4150) connecting to nsqd 2024/05/24 19:56:15 INF 102 (localhost:4150) connecting to nsqd 2024/05/24 19:57:13 INF 103 (localhost:4150) connecting to nsqd 2024/05/24 19:57:14 INF 104 (localhost:4150) connecting to nsqd 2024/05/24 19:57:20 INF 105 (localhost:4150) connecting to nsqd 2024/05/24 19:57:45 INF 106 (localhost:4150) connecting to nsqd 2024/05/24 19:57:46 INF 107 (localhost:4150) connecting to nsqd 2024/05/24 19:57:46 INF 108 (localhost:4150) connecting to nsqd 2024/05/24 20:01:16 INF 109 (localhost:4150) connecting to nsqd 2024/05/24 20:01:35 INF 110 (localhost:4150) connecting to nsqd 2024/05/24 20:01:40 INF 111 (localhost:4150) connecting to nsqd "accessModes": [ "affinity": "affinity": { "affinity": { "podAntiAffinity": { "allowPrivilegeEscalation": false, "allowPrivilegeEscalation": false, "annotations": {"keep-backups":"false","keep-data":"false"}, "apiVersion": "apps/v1", "apiVersion": "batch/v1", "apiVersion": "v1", "apiVersion": "v1", "apiVersion": "v1", "backrest-command": "backup", "backrest-command": "stanza-create", "claimName": "cluster1-pgbr-repo" "cluster1" "command": [ "configMap": { "configMap": { "configMapKeyRef": { "containerName": "database", "containerPort": 10000, "containerPort": 2022, "containerPort": 5432, "containerPort": 8009, "containers": [ "containers": [{ "cpu": "1", "cpu": "200m", "cpu": "500m", "cpu": "500m", "crunchydata" "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "crunchy-pgha-scope": "cluster1","pgo-version": "1.6.0","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin" "defaultMode": 288 "defaultMode": 288, "defaultMode": 288, "defaultMode": 420, "deployment-name": "cluster1" "deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.6.0","name": "cluster1","pg-cluster": "cluster1" "deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.6.0","name": "cluster1","pg-cluster": "cluster1" "deployment-name": "cluster1-repl1" "deployment-name": "cluster1-repl1","pgo-version": "1.6.0","name": "cluster1-replica","pg-cluster": "cluster1" "deployment-name": "cluster1-repl1","pgo-version": "1.6.0","name": "cluster1-replica","pg-cluster": "cluster1" "deployment-name": "cluster1-repl2" "deployment-name": "cluster1-repl2","pgo-version": "1.6.0","name": "cluster1-replica","pg-cluster": "cluster1" "deployment-name": "cluster1-repl2","pgo-version": "9.9.9","name": "cluster1-replica","pg-cluster": "cluster1" "divisor": "1m", "dnsPolicy": "ClusterFirst" "downwardAPI": { "emptyDir": { "emptyDir": { "emptyDir": { "emptyDir": { "env": [ "env": [{ "env": [ { "exec": { "fieldPath": "metadata.annotations" "fieldPath": "metadata.labels" "fieldPath": "metadata.labels['crunchy-pgha-scope']" "fieldPath": "metadata.namespace" "fieldRef": { "fieldRef": { "fsGroup": 2, I0524 18:49:30.727821 1 shared_informer.go:240] Waiting for caches to sync for namespace I0524 18:49:30.828259 1 shared_informer.go:247] Caches are synced for namespace I0524 18:49:30.970727 1 shared_informer.go:240] Waiting for caches to sync for version-service-7914 I0524 18:49:31.071436 1 shared_informer.go:247] Caches are synced for version-service-7914 "image": "perconalab/percona-postgresql-operator:main-ppg12-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg12-pgbadger", "image": "perconalab/percona-postgresql-operator:main-ppg13-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg13-pgbadger", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbadger", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbouncer", "image": "perconalab/percona-postgresql-operator:main-ppg14-postgres-ha", "image": "perconalab/percona-postgresql-operator:PR-776-0087ef129-pgo-rmdata", "image": "percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest", "image": "percona/percona-postgresql-operator:1.6.0-ppg12-pgbackrest-repo", "image": "percona/percona-postgresql-operator:1.6.0-ppg12-pgbadger", "image": "percona/percona-postgresql-operator:1.6.0-ppg12-pgbouncer", "image": "percona/percona-postgresql-operator:1.6.0-ppg12-postgres-ha", "image": "percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest", "image": "percona/percona-postgresql-operator:1.6.0-ppg13-pgbackrest-repo", "image": "percona/percona-postgresql-operator:1.6.0-ppg13-pgbadger", "image": "percona/percona-postgresql-operator:1.6.0-ppg13-pgbouncer", "image": "percona/percona-postgresql-operator:1.6.0-ppg13-postgres-ha", "imagePullPolicy": "IfNotPresent" "imagePullPolicy": "IfNotPresent", "imagePullPolicy": "IfNotPresent", "initialDelaySeconds": 15 "initialDelaySeconds": 20, "initialDelaySeconds": 30, "items": [ "items": [ "items": [ "key": "config", "key": "config", "key": "crunchy-pgbouncer", "key": "init" "key": "password" "key": "pg-cluster", "key": "pgo-backrest", "key": "pgo-backrest-job", "key": "pgo-pg-database", "key": "pg-pod-anti-affinity", "key": "tls.crt", "key": "tls.key", "key": "vendor", "kind": "Deployment", "kind": "Job", "kind": "PersistentVolumeClaim", "kind": "Service", "labels": { "labels": { "labels": { "labelSelector": { "limits": { "livenessProbe": { "matchExpressions": [ "matchLabels": { "matchLabels": { "maxSurge": 1 "maxUnavailable": 1, "medium": "Memory", "medium": "Memory" "medium": "Memory", "medium": "Memory", "medium": "Memory", "memory": "128Mi" "memory": "256Mi" "memory": "48Mi" "memory": "64Mi" "metadata": { "metadata": { "metadata": { "mountPath": "/backrestrepo", "mountPath": "/dev/shm", "mountPath": "/etc/pgbackrest" "mountPath": "/etc/pgbackrest/conf.d", "mountPath": "/etc/podinfo", "mountPath": "/etc/ssh", "mountPath": "/etc/ssh", "mountPath": "/home/pgbackrest" "mountPath": "/pgconf", "mountPath": "/pgconf/", "mountPath": "/pgconf/pgreplicator", "mountPath": "/pgconf/pgsuper", "mountPath": "/pgconf/pguser", "mountPath": "/pgconf/tls", "mountPath": "/pgconf/tls/pgbouncer", "mountPath": "/pgconf/tls-replication", "mountPath": "/pgdata", "mountPath": "/pgdata", "mountPath": "/report", "mountPath": "/sshd", "mountPath": "/sshd", "mountPath": "/tmp", "mountPath": "/tmp", "mountPath": "/tmp" "mountPath": "/tmp", "name": "backrest", "name": "backrest-backup-cluster1", "name": "backrestrepo", "name": "backrestrepo", "name": "BACKREST_SKIP_CREATE_STANZA", "name": "BADGER_TARGET", "name": "cluster1", "name": "cluster1", "name": "cluster1-backrest-shared-repo", "name": "cluster1-backrest-shared-repo" "name": "cluster1-backrest-shared-repo", "name": "cluster1-backrest-shared-repo", "name": "cluster1-backrest-shared-repo", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer-cm" "name": "cluster1-pgbouncer-secret", "name": "cluster1-pgbouncer-secret", "name": "cluster1-pgbr-repo", "name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.6.0" "name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "9.9.9" "name": "cluster1-pgha-config", "name": "cluster1-pgha-config", "name": "cluster1-repl1", "name": "cluster1-repl1", "name": "cluster1-repl2", "name": "cluster1-repl2", "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1","pgo-version": "1.6.0" "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1","pgo-version": "1.6.0" "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.6.0" "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.6.0" "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "9.9.9" "name": "cluster1-replication-ssl-keypair", "name": "cluster1-rmdata-bdzs", "name": "cluster1-rmdata-bdzs", "name": "cluster1-rmdata-ethu", "name": "cluster1-rmdata-ethu", "name": "cluster1-rmdata-gxap", "name": "cluster1-rmdata-gxap", "name": "cluster1-rmdata-ibll", "name": "cluster1-rmdata-ibll", "name": "cluster1-rmdata-ilhp", "name": "cluster1-rmdata-ilhp", "name": "cluster1-rmdata-lbqq", "name": "cluster1-rmdata-lbqq", "name": "cluster1-rmdata-tsps", "name": "cluster1-rmdata-tsps", "name": "cluster1-rmdata-wegj", "name": "cluster1-rmdata-wegj", "name": "cluster1-rmdata-zqjj", "name": "cluster1-rmdata-zqjj", "name": "cluster1-ssl-ca" "name": "cluster1-ssl-ca" "name": "cluster1-ssl-keypair" "name": "cluster1-ssl-keypair" "name": "cluster1-stanza-create", "name": "COMMAND", "name": "COMMAND_OPTS", "name": "database", "name": "dshm" "name": "dshm", "name": "ENABLE_SSHD", "name": "fulldata-cluster1", "name": "IS_BACKUP", "name": "IS_REPLICA", "name": "LD_PRELOAD", "name": "LD_PRELOAD", "name": "MODE", "name": "NAMESPACE", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_PASSWD", "name": "NSS_WRAPPER_PASSWD", "name": "patroni" "name": "PATRONI_KUBERNETES_LABELS", "name": "PATRONI_KUBERNETES_NAMESPACE", "name": "PATRONI_KUBERNETES_SCOPE_LABEL", "name": "PATRONI_LOG_LEVEL", "name": "PATRONI_POSTGRESQL_DATA_DIR", "name": "PATRONI_SCOPE", "name": "pgbackrest" "name": "pgbackrest-conf", "name": "pgbackrest-conf", "name": "pgbackrest-config" "name": "pgbackrest-config", "name": "PGBACKREST_DB_HOST", "name": "PGBACKREST_DB_PATH", "name": "PGBACKREST_DB_PATH", "name": "pgbackrest-home", "name": "pgbackrest-home", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_REPO1_HOST", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_STANZA", "name": "PGBACKREST_STANZA", "name": "pgbadger" "name": "pgbadger", "name": "PGBADGER_SERVICE_PORT", "name": "pgbouncer" "name": "pgbouncer", "name": "pgbouncer-conf", "name": "pgbouncer-conf", "name": "PG_CLUSTER", "name": "pgconf-volume" "name": "pgconf-volume", "name": "pgdata", "name": "pgdata", "name": "pgdata", "name": "PGHA_DATABASE", "name": "PGHA_INIT", "name": "PGHA_PASSWORD_TYPE", "name": "PGHA_PGBACKREST", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_S3_VERIFY_TLS", "name": "PGHA_PG_PORT", "name": "PGHA_REPLICA_REINIT_ON_START_FAIL", "name": "PGHA_SCOPE", "name": "PGHA_STANDBY", "name": "PGHA_SYNC_REPLICATION", "name": "PGHA_TLS_ENABLED", "name": "PGHA_TLS_ONLY", "name": "PGHA_USER", "name": "PGHOST", "name": "PG_PASSWORD", "name": "PG_PRIMARY_SERVICE_NAME", "name": "PITR_TARGET", "name": "podinfo" "name": "podinfo", "name": "PODNAME", "name": "postgres" "name": "primary-volume" "name": "primary-volume", "name": "REMOVE_BACKUP", "name": "REMOVE_DATA", "name": "REPLICA_NAME", "name": "report", "name": "report", "name": "rmdata", "name": "root-volume" "name": "root-volume", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "sshd", "name": "sshd", "name": "sshd", "name": "sshd", "name": "tls-pgbouncer" "name": "tls-pgbouncer", "name": "tls-replication" "name": "tls-replication", "name": "tls-server" "name": "tls-server", "name": "tmp" "name": "tmp" "name": "tmp", "name": "tmp", "name": "tmp", "name": "tmp" "name": "user-volume" "name": "user-volume", "nodePort": 0 "operator": "Exists" "operator": "In", "/opt/crunchy/bin/postgres-ha/health/pgha-liveness.sh" "/opt/crunchy/bin/postgres-ha/health/pgha-readiness.sh" "optional": true "path": "annotations" "path": "cpu_limit", "path": "cpu_request", "path": "labels" "path": "mem_limit", "path": "mem_request", "path": "ssh_config" "path": "ssh_config" "path": "tls-replication.crt" "path": "tls-replication.key" "periodSeconds": 10 "periodSeconds": 15, "persistentVolumeClaim": { "persistentVolumeClaim":{"claimName":"cluster1"} "persistentVolumeClaim":{"claimName":"cluster1-repl1"} "persistentVolumeClaim":{"claimName":"cluster1-repl2"} "pg-cluster": "cluster1" "pg-cluster": "cluster1", "pg-cluster": "cluster1" "pg-cluster": "cluster1", "pg-cluster": "cluster1" "pg-cluster": "cluster1", "pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.6.0","name": "cluster1" "pg-cluster": "cluster1","deployment-name": "cluster1-repl1","pgo-version": "1.6.0","name": "cluster1-replica" "pg-cluster": "cluster1","deployment-name": "cluster1-repl1","pgo-version": "1.6.0","name": "cluster1-replica" "pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.6.0","name": "cluster1-replica" "pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.6.0","name": "cluster1-replica" "pgo-backrest-job": "true", "pgo-backrest-repo": "true" "pgo-backrest-repo": "true", "pgo-backrest-repo": "true" "pgo-backrest": "true", "pgo-pg-database": "true", "pgo-pg-database": "true", "pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.6.0","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1" "pgo-version": "1.6.0","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1" "pgo-version": "1.6.0","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1" "pgo-version": "1.6.0","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pgo-version": "1.6.0","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pgo-version": "1.6.0","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2" "pgo-version": "1.6.0","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2" "pgo-version": "9.9.9","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1" "pgo-version": "9.9.9","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pgo-version": "9.9.9","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pg-pod-anti-affinity": "preferred", "pg-pod-anti-affinity": "preferred", "pgrmdata": "true", "pgrmdata": "true", "podAffinityTerm": { "podAntiAffinity": { "podAntiAffinity": { "podAntiAffinity": { "port": 10000 "port": 2022, "ports": [{ "ports": [{ "ports": [ { "preferredDuringSchedulingIgnoredDuringExecution": [ "privileged": false, "privileged": false, "projected": { "projected": { "projected": { "projected": { "sources": [] } "protocol": "TCP", "protocol": "TCP", "protocol": "TCP", "readinessProbe": { "readinessProbe": { "readOnly": false "readOnly": false "readOnly": false "readOnly": false "readOnlyRootFilesystem": true "readOnlyRootFilesystem": true "readOnly": true "readOnly": true "readOnly": true "readOnly": true "ReadWriteOnce" "replicas": 0, "replicas": 1, "replicas": 3, "requests": { "requests": { "resourceFieldRef": { "resource": "limits.cpu" "resource": "limits.memory" "resource": "requests.cpu" "resource": "requests.memory" "resources": { "resources": { "resources": { "restartPolicy": "Always", "restartPolicy": "Never" "rollingUpdate": { "runAsNonRoot": true "secret": { "secret": { "secret": { "secret": { "secret": { "secretKeyRef": { "secretName": "cluster1-backrest-repo-config" "secretName": "cluster1-backrest-repo-config", "secretName": "cluster1-backrest-repo-config" "secretName": "cluster1-backrest-repo-config", "secretName": "cluster1-postgres-secret" "secretName": "cluster1-primaryuser-secret" "secretName": "cluster1-some-name-secret" "securityContext": { "securityContext": { "securityContext": {}, "securityContext": { "securityContext": {"fsGroup": 26,"supplementalGroups": [1001]}, "securityContext": {"runAsNonRoot":true,"fsGroup":26}, "selector": { "selector": { "serviceAccountName": "pgo-backrest", "serviceAccountName": "pgo-default", "serviceAccountName": "pgo-pg", "serviceAccountName": "pgo-target", "service-name": "cluster1-backrest-shared-repo", "service-name": "cluster1-backrest-shared-repo", "service-name": "cluster1-pgbouncer", "service-name": "cluster1-pgbouncer", "service-name": "cluster1-pgbouncer", "sessionAffinity": "None" "sizeLimit": "128Ki" "sizeLimit": "16Mi" "sizeLimit": "1Mi" "sizeLimit": "2Mi" "sizeLimit": "64Mi" "sizeLimit": "64Mi" "sources": [ "sources": [ "sources": [ "spec": { "spec": { "spec": { "storage": "1Gi" "strategy": { "targetPort": 2022, "tcpSocket": { "template": { time="2024-05-24T18:49:30Z" level=warning msg="PGAdminStorage setting not set, will use PrimaryStorage setting" func="github.com/percona/percona-postgresql-operator/internal/config.(*PgoConfig).Validate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/config/pgoconfig.go:337" version=1.7.0 time="2024-05-24T18:49:57Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T18:50:03Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T18:50:43Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T18:50:43Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T18:51:19Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T18:51:19Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T18:51:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T18:51:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T18:51:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T18:58:57Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.7.0 time="2024-05-24T18:59:08Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T18:59:08Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl1/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.7.0 time="2024-05-24T19:01:03Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:04Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:05Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:06Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:07Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:11Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:15Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:16Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:19Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:21Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:22Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:01:28Z" level=error msg="jobs.batch \"cluster1-rmdata-zqjj\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:01:47Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:01:53Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:02:34Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:02:34Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:03:09Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:03:10Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:03:10Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:04:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:04:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:04:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:05:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:05:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:05:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:07:05Z" level=error msg="command terminated with exit code 2" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:07:07Z" level=error msg="update perconapgcluster: check is pgcluster primary: no role labels in pod cluster1-f479494fc-kfmj7" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:532" version=1.7.0 time="2024-05-24T19:07:07Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.7.0 time="2024-05-24T19:07:21Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:07:21Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl1/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.7.0 time="2024-05-24T19:09:10Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:12Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:13Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:14Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:15Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:21Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:25Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:26Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:30Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:33Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:34Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:09:39Z" level=error msg="jobs.batch \"cluster1-rmdata-tsps\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:09:54Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:09:59Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:10:38Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:10:38Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:11:14Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:11:14Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:14:20Z" level=error msg="14-recommended value for spec.upgradeOptions.apply option is not supported" func="github.com/percona/percona-postgresql-operator/percona/controllers/version.EnsureVersion()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/version/version.go:48" version=1.7.0 time="2024-05-24T19:14:20Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.7.0 time="2024-05-24T19:14:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:14:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:14:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:15:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:15:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:15:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:17:21Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:23Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:24Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:25Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:28Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:33Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:38Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:40Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:41Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:17:46Z" level=error msg="jobs.batch \"cluster1-rmdata-lbqq\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:18:03Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:18:09Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:18:48Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:18:48Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:19:23Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:19:23Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:19:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T19:19:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:19:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T19:21:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:21:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:21:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:22:56Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.7.0 time="2024-05-24T19:23:06Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:23:06Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl1/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.7.0 time="2024-05-24T19:25:00Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:02Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:03Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:04Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:05Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:07Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:12Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:13Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:16Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:18Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:19Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:25:24Z" level=error msg="jobs.batch \"cluster1-rmdata-ethu\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:25:45Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:25:50Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:26:27Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:26:27Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:27:03Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:27:04Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:27:04Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:28:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:28:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:28:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:29:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:29:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:29:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:30:59Z" level=error msg="command terminated with exit code 2" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:31:00Z" level=error msg="update perconapgcluster: check is pgcluster primary: no role labels in pod cluster1-f479494fc-m4jrn" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:532" version=1.7.0 time="2024-05-24T19:31:00Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.7.0 time="2024-05-24T19:31:14Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:31:14Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl2/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.7.0 time="2024-05-24T19:33:17Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:18Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:19Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:20Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:21Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:22Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:25Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:29Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:32Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:35Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:33:40Z" level=error msg="jobs.batch \"cluster1-rmdata-gxap\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:34:03Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:34:08Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:34:48Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:34:48Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:35:22Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:35:22Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:35:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T19:35:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:35:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T19:39:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:39:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:39:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:40:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:40:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:40:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:41:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:41:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:41:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:42:09Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.7.0 time="2024-05-24T19:42:19Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:42:19Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl1/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.7.0 time="2024-05-24T19:44:14Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:16Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:17Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:18Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:19Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:21Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:25Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:26Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:29Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:30Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:31Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:44:36Z" level=error msg="jobs.batch \"cluster1-rmdata-bdzs\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:44:58Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:45:03Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:45:41Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:45:41Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:46:16Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:46:17Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:46:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T19:46:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:46:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T19:46:58Z" level=error msg="update perconapgcluster: ensure version failed to check version: version service apply: Get \"http://not-a-version-service/versions/v1/pg-operator/9.9.9/latest?customResourceUid=927d60dd-58ae-4b1f-8121-fef2f3f1fee8&databaseVersion=13.14+-+Percona+Distribution\": dial tcp: lookup not-a-version-service on 10.113.0.10:53: no such host" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:501" version=1.7.0 time="2024-05-24T19:46:58Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.7.0 time="2024-05-24T19:47:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T19:47:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:47:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T19:50:00Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:02Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:03Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:04Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:05Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:06Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:11Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:15Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:16Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:20Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:22Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:23Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:50:28Z" level=error msg="jobs.batch \"cluster1-rmdata-ibll\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:50:44Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:50:49Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:51:27Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:51:27Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:52:01Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:52:02Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:52:02Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:53:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:53:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:53:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:54:31Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.7.0 time="2024-05-24T19:54:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T19:54:31Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.7.0 time="2024-05-24T19:55:37Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:39Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:40Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:41Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:44Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:48Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:49Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:53Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:55Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:55:56Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T19:56:01Z" level=error msg="jobs.batch \"cluster1-rmdata-ilhp\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T19:56:19Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:56:24Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T19:57:13Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T19:57:13Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T19:57:46Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T19:57:46Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T20:01:17Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:18Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:19Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:20Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:23Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:28Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:29Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:32Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:34Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:35Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:01:40Z" level=error msg="jobs.batch \"cluster1-rmdata-wegj\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 "timeoutSeconds": 10 "topologyKey": "kubernetes.io/hostname" "type": "ClusterIP", "type": "RollingUpdate", "value": "" "value": "10000" "value": "5432" "value": "5432" "value": "/backrestrepo/cluster1-backrest-shared-repo" "value": "/backrestrepo/cluster1-backrest-shared-repo" "value": "backup" "value": "cluster1" "value": "cluster1" "value": "cluster1-backrest-shared-repo" "value": "cluster1-backrest-shared-repo-574b598fb6-df4tq" "value": "cluster1-backrest-shared-repo-574b598fb6-ndlhc" "value": "cluster1-backrest-shared-repo-574b598fb6-tzkkt" "value": "cluster1-backrest-shared-repo-574b598fb6-xlvl6" "value": "cluster1-backrest-shared-repo-57f668d478-258cw" "value": "cluster1-backrest-shared-repo-57f668d478-59n6c" "value": "cluster1-backrest-shared-repo-5f647d77b5-2b5rr" "value": "cluster1-backrest-shared-repo-5f647d77b5-9t6dj" "value": "cluster1-backrest-shared-repo-5f647d77b5-dxbk9" "value": "cluster1-backrest-shared-repo-9f8cd5549-8sb26" "value": "cluster1-backrest-shared-repo-9f8cd5549-9hcwt" "value": "cluster1-backrest-shared-repo-9f8cd5549-fvn6h" "value": "cluster1-backrest-shared-repo-9f8cd5549-g7dzf" "value": "cluster1-backrest-shared-repo-cbc58f67f-4b5f7" "value": "cluster1-repl1" "value": "cluster1-repl2" "value": "crunchy-pgha-scope" "value": "db" "value": "db" "value": " --db-host=10.59.25.47 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.53 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.59 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.64 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.70 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.75 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.81 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.86 --db-path=/pgdata/cluster1" "value": " --db-host=10.59.25.91 --db-path=/pgdata/cluster1" "value": "false" "value": "false" "valueFrom": { "value": "INFO" "value": "pgbackrest" "value": "pgbackrest-repo" "value": "/pgdata/cluster1" "value": "/pgdata/cluster1" "value": "/pgdata/cluster1-repl1" "value": "/pgdata/cluster1-repl1" "value": "/pgdata/cluster1-repl2" "value": "/pgdata/cluster1-repl2" "value": "posix" "value": "posix" "value": "postgres" "values": [ "value": "stanza-create" "value": "/tmp" "value": "/tmp" "value": "/tmp/nss_wrapper/pgbackrest-repo/group" "value": "/tmp/nss_wrapper/pgbackrest-repo/passwd" "value": "/tmp/nss_wrapper/postgres/group" "value": "/tmp/nss_wrapper/postgres/passwd" "value": "true" "value": "true" "value": "--type=full --db-host=10.59.24.63 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.24.71 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.24.84 --db-path=/pgdata/cluster1-repl2" "value": "--type=full --db-host=10.59.24.93 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.47 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.53 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.59 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.64 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.70 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.75 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.77 --db-path=/pgdata/cluster1-repl2" "value": "--type=full --db-host=10.59.25.81 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.86 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.25.88 --db-path=/pgdata/cluster1-repl2" "value": "--type=full --db-host=10.59.25.91 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.59.26.45 --db-path=/pgdata/cluster1-repl1" "value": "--type=full --db-host=10.59.26.60 --db-path=/pgdata/cluster1-repl1" "value": "/usr/lib64/libnss_wrapper.so" "value": "/usr/lib64/libnss_wrapper.so" "value": "{vendor: \"crunchydata\"}" "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "volumeMounts": [ "volumeMounts": [{ "volumeMounts": [ "volumes": [ "volumes": [{ "weight": 1, + destroy_operator + kubectl_bin delete -f /tmp/tmp.IoigtEzcpK/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.81QQLv7jB9 ++ mktemp + local LAST_ERR=/tmp/tmp.1JWF8hR01G + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete -f /tmp/tmp.IoigtEzcpK/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete -f /tmp/tmp.IoigtEzcpK/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete -f /tmp/tmp.IoigtEzcpK/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.81QQLv7jB9 + cat /tmp/tmp.1JWF8hR01G Error from server (NotFound): error when deleting "/tmp/tmp.IoigtEzcpK/operator.yaml": serviceaccounts "pgo-deployer-sa" not found Error from server (NotFound): error when deleting "/tmp/tmp.IoigtEzcpK/operator.yaml": clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): error when deleting "/tmp/tmp.IoigtEzcpK/operator.yaml": configmaps "pgo-deployer-cm" not found Error from server (NotFound): error when deleting "/tmp/tmp.IoigtEzcpK/operator.yaml": clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-crb" not found Error from server (NotFound): error when deleting "/tmp/tmp.IoigtEzcpK/operator.yaml": jobs.batch "pgo-deploy" not found + rm /tmp/tmp.81QQLv7jB9 /tmp/tmp.1JWF8hR01G + return 1 + true + yq w -d4 /tmp/tmp.IoigtEzcpK/operator.yaml 'spec.template.spec.containers[0].env[0].value' uninstall + kubectl_bin apply -f /tmp/tmp.IoigtEzcpK/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.lUlgCfq9y4 ++ mktemp + local LAST_ERR=/tmp/tmp.eomQt03bje + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl apply -f /tmp/tmp.IoigtEzcpK/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.lUlgCfq9y4 serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb created job.batch/pgo-deploy created + cat /tmp/tmp.eomQt03bje + rm /tmp/tmp.lUlgCfq9y4 /tmp/tmp.eomQt03bje + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=version-service-7914 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-7914 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.IoigtEzcpK/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.gngdbrtuZ2 ++ mktemp + local LAST_ERR=/tmp/tmp.l9F3C8NvCO + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete -f /tmp/tmp.IoigtEzcpK/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gngdbrtuZ2 serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.l9F3C8NvCO + rm /tmp/tmp.gngdbrtuZ2 /tmp/tmp.l9F3C8NvCO + return 0 + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + awk '{print$1}' ++ mktemp + xargs kubectl delete ns --grace-period=0 --force=true + local LAST_OUT=/tmp/tmp.m6l374EG4P ++ mktemp + local LAST_ERR=/tmp/tmp.1o8MqulWg9 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.m6l374EG4P + cat /tmp/tmp.1o8MqulWg9 + rm /tmp/tmp.m6l374EG4P /tmp/tmp.1o8MqulWg9 + return 0 Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. namespace "version-service-7914" force deleted + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.skHVGk4tDH ++ mktemp + local LAST_ERR=/tmp/tmp.KbPEODdpWB + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.skHVGk4tDH + cat /tmp/tmp.KbPEODdpWB Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.skHVGk4tDH /tmp/tmp.KbPEODdpWB + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.HF9XXVooLh ++ mktemp + local LAST_ERR=/tmp/tmp.n1DNLjUneI + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.HF9XXVooLh + cat /tmp/tmp.n1DNLjUneI Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.HF9XXVooLh /tmp/tmp.n1DNLjUneI + return 1 + true + rm -rf /tmp/tmp.IoigtEzcpK