++ echo 'Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/logs/version-service.log' Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/logs/version-service.log ++ '[' -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP= ++ oc get projects error: the server doesn't have a resource type "projects" +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep '\-eks\-' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ '[' ']' +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ kubectl version -o json +++ /usr/bin/sed -r 's/[^0-9.]+//g' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ KUBE_VERSION=1.26 +++ helm version -c +++ /usr/bin/sed -re 's/.*SemVer:"([^"]+)".*/\1/; s/.*\bVersion:"([^"]+)".*/\1/' ++ HELM_VERSION=v3.14.4 ++ '[' v3 == v2 ']' + VS_URL=http://version-service + VS_PORT=11000 + VS_ENDPOINT=http://version-service:11000 + main + create_namespace version-service-20560 + local namespace=version-service-20560 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + cleanup_rbac + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + xargs kubectl delete ns + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + kubectl_bin get ns ++ mktemp ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.5WngNsRsJ6 + local LAST_OUT=/tmp/tmp.s3pV9ttGzh ++ mktemp + local LAST_ERR=/tmp/tmp.frMSvFrrst + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp + local LAST_ERR=/tmp/tmp.JbbiPHyrLI + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.5WngNsRsJ6 + cat /tmp/tmp.JbbiPHyrLI E0424 16:25:07.165380 17056 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:07.488377 17056 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:07.595177 17056 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:07.701005 17056 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request + rm /tmp/tmp.5WngNsRsJ6 /tmp/tmp.JbbiPHyrLI + return 0 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level E0424 16:25:09.334998 17286 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:09.454246 17286 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:09.561005 17286 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:09.667687 17286 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: resource(s) were provided, but no name was specified + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.s3pV9ttGzh + cat /tmp/tmp.frMSvFrrst E0424 16:25:15.329119 17614 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:15.444106 17614 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:15.549150 17614 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:15.654403 17614 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.s3pV9ttGzh /tmp/tmp.frMSvFrrst + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp + local LAST_OUT=/tmp/tmp.K8t6tPdVAn ++ mktemp + local LAST_ERR=/tmp/tmp.6NgzNeYN16 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.K8t6tPdVAn + cat /tmp/tmp.6NgzNeYN16 E0424 16:25:33.326451 18325 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:33.547065 18325 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:33.654928 18325 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:33.761673 18325 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.K8t6tPdVAn /tmp/tmp.6NgzNeYN16 + return 1 + true + kubectl delete MutatingWebhookConfiguration/chaos-mesh-mutation ValidatingWebhookConfiguration/chaos-mesh-validation ValidatingWebhookConfiguration/validate-auth E0424 16:25:43.588938 18685 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:43.903749 18685 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:44.010042 18685 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0424 16:25:44.116844 18685 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request Error from server (NotFound): mutatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-mutation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-validation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "validate-auth" not found + true + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin delete namespace version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.fpBmaLD8ds ++ mktemp + local LAST_ERR=/tmp/tmp.iECu3cjNLx + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete namespace version-service-20560 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete namespace version-service-20560 + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete namespace version-service-20560 + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.fpBmaLD8ds + cat /tmp/tmp.iECu3cjNLx Error from server (NotFound): namespaces "version-service-20560" not found + rm /tmp/tmp.fpBmaLD8ds /tmp/tmp.iECu3cjNLx + return 1 + : + wait_for_delete namespace/version-service-20560 + local res=namespace/version-service-20560 + set +o xtrace namespace/version-service-20560 - Error from server (NotFound): namespaces "version-service-20560" not found + kubectl_bin create namespace version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.xIzvdNmdkg ++ mktemp + local LAST_ERR=/tmp/tmp.H1pG1mVMv1 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl create namespace version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xIzvdNmdkg namespace/version-service-20560 created + cat /tmp/tmp.H1pG1mVMv1 + rm /tmp/tmp.xIzvdNmdkg /tmp/tmp.H1pG1mVMv1 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.1xMGilI1np +++ mktemp ++ local LAST_ERR=/tmp/tmp.R4JU183Jra ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.1xMGilI1np ++ cat /tmp/tmp.R4JU183Jra ++ rm /tmp/tmp.1xMGilI1np /tmp/tmp.R4JU183Jra ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-723-7ed19e708-2-version-service --namespace=version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.zy4onB8VLR ++ mktemp + local LAST_ERR=/tmp/tmp.1k0UGuUwn5 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-723-7ed19e708-2-version-service --namespace=version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.zy4onB8VLR Context "gke_cloud-dev-112233_us-central1-c_jen-pg-723-7ed19e708-2-version-service" modified. + cat /tmp/tmp.1k0UGuUwn5 + rm /tmp/tmp.zy4onB8VLR /tmp/tmp.1k0UGuUwn5 + return 0 + deploy_operator '' /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/operator.yaml disabled install version-service-20560 false http://version-service:11000 + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/operator.yaml + local namespace_mode=disabled + local action=install + local namespace=version-service-20560 + local dont_send_telemetry=false + local version_service_uri_override=http://version-service:11000 + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + prepare_operator_yaml '' /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/operator.yaml disabled install version-service-20560 false + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/operator.yaml + local namespace_mode=disabled + local operator_action=install + local namespace=version-service-20560 + local dont_send_telemetry=false + yq r -d2 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/operator.yaml 'data[values.yaml]' + /usr/bin/sed -e 's#pgo_operator_namespace: .*#pgo_operator_namespace: "version-service-20560"#g' + /usr/bin/sed -e 's#^disable_telemetry: .*#disable_telemetry: "false"#g' + /usr/bin/sed -e 's#^namespace: .*#namespace: "version-service-20560"#g' + cat /dev/fd/63 - ++ echo 'crunchy_debug: "true"' ++ cut -d: -f1 ++ echo perconalab/percona-postgresql-operator:PR-723-7ed19e708 + /usr/bin/sed -e 's#namespace_mode: .*#namespace_mode: "disabled"#g' + /usr/bin/sed -e 's#pgo_image_prefix: .*#pgo_image_prefix: "perconalab/percona-postgresql-operator"#g' ++ echo perconalab/percona-postgresql-operator:PR-723-7ed19e708 ++ cut -d: -f2 + /usr/bin/sed -e 's#pgo_image_tag: .*#pgo_image_tag: "PR-723-7ed19e708"#g' + [[ -n '' ]] + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + yq w '-d*' /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/operator.yaml metadata.namespace version-service-20560 + yq w -d4 - 'spec.template.spec.containers[0].image' perconalab/percona-postgresql-operator:PR-723-7ed19e708-pgo-deployer + yq w -d3 - 'subjects[0].namespace' version-service-20560 + yq w -d4 - 'spec.template.spec.containers[0].env[0].value' install ++ cat /tmp/tmp.09n2hxddFJ/operator.ini + yq w -d2 - -d2 'data[values.yaml]' 'crunchy_debug: "true" archive_mode: "true" archive_timeout: "60" ccp_image_pull_secret: "" ccp_image_pull_secret_manifest: "" create_rbac: "true" delete_operator_namespace: "false" delete_watched_namespaces: "false" disable_telemetry: "false" namespace: "version-service-20560" namespace_mode: "disabled" pgo_image_prefix: "perconalab/percona-postgresql-operator" pgo_image_pull_policy: "Always" pgo_image_pull_secret: "" pgo_image_pull_secret_manifest: "" pgo_image_tag: "PR-723-7ed19e708" pgo_operator_namespace: "version-service-20560"' + [[ -n '' ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.PG6Jojw3GR ++ mktemp + local LAST_ERR=/tmp/tmp.qfsSGo2PFA + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PG6Jojw3GR serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb created job.batch/pgo-deploy created + cat /tmp/tmp.qfsSGo2PFA + rm /tmp/tmp.PG6Jojw3GR /tmp/tmp.qfsSGo2PFA + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 15 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 16 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 17 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 18 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 19 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 20 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 21 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 22 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 23 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 24 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 25 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 26 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 27 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 28 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.09n2hxddFJ/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.vsIPj7lr5A ++ mktemp + local LAST_ERR=/tmp/tmp.zqT2f1vYaF + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete -f /tmp/tmp.09n2hxddFJ/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.vsIPj7lr5A serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.zqT2f1vYaF + rm /tmp/tmp.vsIPj7lr5A /tmp/tmp.zqT2f1vYaF + return 0 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gZmqX5hZqa +++ mktemp ++ local LAST_ERR=/tmp/tmp.oLM5nWLOtk ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.gZmqX5hZqa ++ cat /tmp/tmp.oLM5nWLOtk ++ rm /tmp/tmp.gZmqX5hZqa /tmp/tmp.oLM5nWLOtk ++ return 0 + wait_pod postgres-operator-b7dcc5b9c-vtq4r + local pod=postgres-operator-b7dcc5b9c-vtq4r + local ns= + set +o xtrace postgres-operator-b7dcc5b9c-vtq4rtrue + [[ -n http://version-service:11000 ]] + kubectl_bin -n version-service-20560 patch deployment/postgres-operator --type json '-p=[{"op":"add","path":"/spec/template/spec/containers/1/env/-","value":{"name":"PERCONA_VS_FALLBACK_URI","value":"http://version-service:11000"}}]' ++ mktemp + local LAST_OUT=/tmp/tmp.4cUoFIdgdA ++ mktemp + local LAST_ERR=/tmp/tmp.jwIBDTxBRn + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 patch deployment/postgres-operator --type json '-p=[{"op":"add","path":"/spec/template/spec/containers/1/env/-","value":{"name":"PERCONA_VS_FALLBACK_URI","value":"http://version-service:11000"}}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.4cUoFIdgdA deployment.apps/postgres-operator patched + cat /tmp/tmp.jwIBDTxBRn + rm /tmp/tmp.4cUoFIdgdA /tmp/tmp.jwIBDTxBRn + return 0 + sleep 15 + deploy_version_service + desc 'install version service' + set +o xtrace ----------------------------------------------------------------------------------- install version service ----------------------------------------------------------------------------------- + kubectl_bin create configmap versions --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.dep.json --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.json ++ mktemp + local LAST_OUT=/tmp/tmp.ePMOvG1q7f ++ mktemp + local LAST_ERR=/tmp/tmp.72d8I7xa8d + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl create configmap versions --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.dep.json --from-file /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/operator.9.9.9.pg-operator.json + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ePMOvG1q7f configmap/versions created + cat /tmp/tmp.72d8I7xa8d + rm /tmp/tmp.ePMOvG1q7f /tmp/tmp.72d8I7xa8d + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/vs.yml ++ mktemp + local LAST_OUT=/tmp/tmp.xnDYre09YK ++ mktemp + local LAST_ERR=/tmp/tmp.LJWDuGVi6F + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/vs.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xnDYre09YK deployment.apps/version-service created service/version-service created + cat /tmp/tmp.LJWDuGVi6F + rm /tmp/tmp.xnDYre09YK /tmp/tmp.LJWDuGVi6F + return 0 + kubectl_bin apply -f - + /usr/bin/sed s/version-service/version-service-cr/g /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/vs.yml ++ mktemp + local LAST_OUT=/tmp/tmp.wlxoDLCHRB ++ yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/vs.yml 'spec.template.spec.containers[0].image' ++ mktemp + yq w - -d0 'spec.template.spec.containers[0].image' perconalab/version-service:main-ea25e30 + local LAST_ERR=/tmp/tmp.UbhvpWNAHw + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.wlxoDLCHRB deployment.apps/version-service-cr created service/version-service-cr created + cat /tmp/tmp.UbhvpWNAHw + rm /tmp/tmp.wlxoDLCHRB /tmp/tmp.UbhvpWNAHw + return 0 + sleep 10 + check_telemetry_transfer http://version-service-cr:11000 disabled enabled + local cr_vs_uri=http://version-service-cr:11000 + local cr_vs_channel=disabled + local telemetry_state=enabled + local cluster=minimal + yq w - spec.upgradeOptions.versionServiceEndpoint http://version-service-cr:11000 + yq r /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/cr-minimal.yaml + yq w - spec.upgradeOptions.apply disabled + spinup_pgcluster minimal /tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml + local cluster=minimal + local config=/tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets minimal version-service-20560 + local cluster=minimal + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name minimal-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.BPSzB05Exq ++ mktemp + local LAST_ERR=/tmp/tmp.VDwvzApgno + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.BPSzB05Exq secret/minimal-users-secret created + cat /tmp/tmp.VDwvzApgno + rm /tmp/tmp.BPSzB05Exq /tmp/tmp.VDwvzApgno + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + kubectl_bin -n version-service-20560 apply -f - + yq w - metadata.name minimal-pmm-secret ++ mktemp + local LAST_OUT=/tmp/tmp.nfMx4XOmeW ++ mktemp + local LAST_ERR=/tmp/tmp.on5d8TapJf + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nfMx4XOmeW secret/minimal-pmm-secret created + cat /tmp/tmp.on5d8TapJf + rm /tmp/tmp.nfMx4XOmeW /tmp/tmp.on5d8TapJf + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.XnCdEeEG73 ++ mktemp + local LAST_ERR=/tmp/tmp.e6fEPdqsv0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.XnCdEeEG73 deployment.apps/pg-client created + cat /tmp/tmp.e6fEPdqsv0 + rm /tmp/tmp.XnCdEeEG73 /tmp/tmp.e6fEPdqsv0 + return 0 + apply_cluster /tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml minimal false false false false false false false false + local path=/tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml + local name=minimal + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w - metadata.labels.crunchy-pgha-scope minimal + yq w - spec.clustername minimal + yq w - spec.database minimal + yq w - spec.name minimal + yq w - metadata.name minimal + yq w - metadata.labels.pg-cluster minimal + yq w - metadata.labels.name minimal + yq w /tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml metadata.annotations.current-primary minimal + yq w - spec.keepData false + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - metadata.labels.deployment-name minimal + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.keepBackups false + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.user some-name + yq w - spec.secretsName minimal-users-secret + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.fkopp81s3e ++ mktemp + local LAST_ERR=/tmp/tmp.qQiiOVE15Q + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.fkopp81s3e perconapgcluster.pg.percona.com/minimal created + cat /tmp/tmp.qQiiOVE15Q + rm /tmp/tmp.fkopp81s3e /tmp/tmp.qQiiOVE15Q + return 0 + wait_deployment minimal-backrest-shared-repo + local name=minimal-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace minimal-backrest-shared-repo... + wait_deployment minimal + local name=minimal + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace minimal.......... + [[ false == \g\c\s\+* ]] + wait_job_completion minimal-stanza-create false + local job=minimal-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' minimal-stanza-create+ retry=0 ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-minimal + local job=backrest-backup-minimal + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' backrest-backup-minimal+ retry=0 ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency minimal + cluster_name=minimal + wait_cluster_status minimal 'pgcluster Initialized' + local cluster_name=minimal + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C5mPsfqal9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9J748JH7NT ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.C5mPsfqal9 ++ cat /tmp/tmp.9J748JH7NT ++ rm /tmp/tmp.C5mPsfqal9 /tmp/tmp.9J748JH7NT ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.slSX8XcxKX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.rBZVAmgG9s +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.slSX8XcxKX +++ cat /tmp/tmp.rBZVAmgG9s +++ rm /tmp/tmp.slSX8XcxKX /tmp/tmp.rBZVAmgG9s +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.5q9GS2zQZ4 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wXBFuWk3dE +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.5q9GS2zQZ4 +++ cat /tmp/tmp.wXBFuWk3dE +++ rm /tmp/tmp.5q9GS2zQZ4 /tmp/tmp.wXBFuWk3dE +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Mi19Ihugz1 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.7Sf7xIoK4D +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Mi19Ihugz1 +++ cat /tmp/tmp.7Sf7xIoK4D +++ rm /tmp/tmp.Mi19Ihugz1 /tmp/tmp.7Sf7xIoK4D +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.HaXNK9JBHA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.RtktR4FauL +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.HaXNK9JBHA +++ cat /tmp/tmp.RtktR4FauL +++ rm /tmp/tmp.HaXNK9JBHA /tmp/tmp.RtktR4FauL +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ajG3S2FHS2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9OaLbDSg5L +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ajG3S2FHS2 +++ cat /tmp/tmp.9OaLbDSg5L +++ rm /tmp/tmp.ajG3S2FHS2 /tmp/tmp.9OaLbDSg5L +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.RIC4aEFe9Y +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.3hEtmOhyPP ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.RIC4aEFe9Y ++++ cat /tmp/tmp.3hEtmOhyPP ++++ rm /tmp/tmp.RIC4aEFe9Y /tmp/tmp.3hEtmOhyPP ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.kc6uUHIXEc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Vo0SRb1o9M +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.kc6uUHIXEc +++ cat /tmp/tmp.Vo0SRb1o9M +++ rm /tmp/tmp.kc6uUHIXEc /tmp/tmp.Vo0SRb1o9M +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@minimal.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FuuGkq7qBH +++ mktemp ++ local LAST_ERR=/tmp/tmp.ICrGbOTyY4 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.FuuGkq7qBH ++ cat /tmp/tmp.ICrGbOTyY4 ++ rm /tmp/tmp.FuuGkq7qBH /tmp/tmp.ICrGbOTyY4 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.O73VhkI0o9 ++ mktemp + local LAST_ERR=/tmp/tmp.QR5fw11Feg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.O73VhkI0o9 + cat /tmp/tmp.QR5fw11Feg + rm /tmp/tmp.O73VhkI0o9 /tmp/tmp.QR5fw11Feg + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pTdsv61DzY ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Tex8IKs3Ki +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.pTdsv61DzY +++ cat /tmp/tmp.Tex8IKs3Ki +++ rm /tmp/tmp.pTdsv61DzY /tmp/tmp.Tex8IKs3Ki +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ej91ZENwsH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CfYKAxJOkU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ej91ZENwsH +++ cat /tmp/tmp.CfYKAxJOkU +++ rm /tmp/tmp.ej91ZENwsH /tmp/tmp.CfYKAxJOkU +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZYQsuJNBKc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.M6ZHrtPfYo +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZYQsuJNBKc +++ cat /tmp/tmp.M6ZHrtPfYo +++ rm /tmp/tmp.ZYQsuJNBKc /tmp/tmp.M6ZHrtPfYo +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.pBWWXSyVgP +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.zkz94hf2M5 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.pBWWXSyVgP ++++ cat /tmp/tmp.zkz94hf2M5 ++++ rm /tmp/tmp.pBWWXSyVgP /tmp/tmp.zkz94hf2M5 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dtFr74F2ms ++++ mktemp +++ local LAST_ERR=/tmp/tmp.QflyzSoXK2 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.dtFr74F2ms +++ cat /tmp/tmp.QflyzSoXK2 +++ rm /tmp/tmp.dtFr74F2ms /tmp/tmp.QflyzSoXK2 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@minimal.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2eBXdAqBzA +++ mktemp ++ local LAST_ERR=/tmp/tmp.H3bb1D4oz4 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.2eBXdAqBzA ++ cat /tmp/tmp.H3bb1D4oz4 ++ rm /tmp/tmp.2eBXdAqBzA /tmp/tmp.H3bb1D4oz4 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.XlqmXPHmMn ++ mktemp + local LAST_ERR=/tmp/tmp.YhqENKHiib + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.XlqmXPHmMn + cat /tmp/tmp.YhqENKHiib + rm /tmp/tmp.XlqmXPHmMn /tmp/tmp.YhqENKHiib + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.68pYrF0ajE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VnJDu8YOvj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.68pYrF0ajE +++ cat /tmp/tmp.VnJDu8YOvj +++ rm /tmp/tmp.68pYrF0ajE /tmp/tmp.VnJDu8YOvj +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.x3a9vllMWB ++++ mktemp +++ local LAST_ERR=/tmp/tmp.HDOK5MdzNu +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.x3a9vllMWB +++ cat /tmp/tmp.HDOK5MdzNu +++ rm /tmp/tmp.x3a9vllMWB /tmp/tmp.HDOK5MdzNu +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.UC7JcsTdxT ++++ mktemp +++ local LAST_ERR=/tmp/tmp.snZ5UOGnxZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.UC7JcsTdxT +++ cat /tmp/tmp.snZ5UOGnxZ +++ rm /tmp/tmp.UC7JcsTdxT /tmp/tmp.snZ5UOGnxZ +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.zqZdrInoLB +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.pATqJjbJrA ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.zqZdrInoLB ++++ cat /tmp/tmp.pATqJjbJrA ++++ rm /tmp/tmp.zqZdrInoLB /tmp/tmp.pATqJjbJrA ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.iSvx7wJrPO ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BC9M9r4cso +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.iSvx7wJrPO +++ cat /tmp/tmp.BC9M9r4cso +++ rm /tmp/tmp.iSvx7wJrPO /tmp/tmp.BC9M9r4cso +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@minimal.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HifBDpzDuU +++ mktemp ++ local LAST_ERR=/tmp/tmp.x3iGso2N1h ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.HifBDpzDuU ++ cat /tmp/tmp.x3iGso2N1h ++ rm /tmp/tmp.HifBDpzDuU /tmp/tmp.x3iGso2N1h ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.9pKrgaSSJL ++ mktemp + local LAST_ERR=/tmp/tmp.sv8wUa29zn + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9pKrgaSSJL + cat /tmp/tmp.sv8wUa29zn + rm /tmp/tmp.9pKrgaSSJL /tmp/tmp.sv8wUa29zn + return 0 + sleep 10 + grep -Eo '\{.*\}' ++ kubectl get pods --selector=run=version-service-cr -o 'jsonpath={.items[0].metadata.name}' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.request.content".msg.customResourceUid)' + grep -E 'server request payload|unary call' + jq 'del(."grpc.start_time")' + kubectl_bin logs version-service-cr-76cc4777bf-lntq9 ++ mktemp + local LAST_OUT=/tmp/tmp.FGN7UOrTnu ++ mktemp + local LAST_ERR=/tmp/tmp.xGFh8anuXx + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl logs version-service-cr-76cc4777bf-lntq9 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FGN7UOrTnu + cat /tmp/tmp.xGFh8anuXx + rm /tmp/tmp.FGN7UOrTnu /tmp/tmp.xGFh8anuXx + return 0 + grep -Eo '\{.*\}' + jq 'del(."grpc.start_time")' + jq 'del(."grpc.request.content".msg.customResourceUid)' ++ kubectl get pods --selector=run=version-service -o 'jsonpath={.items[0].metadata.name}' + jq 'del(."grpc.time_ms")' + grep -E 'server request payload|unary call' + kubectl_bin logs version-service-5cd4db57-jgs4f ++ mktemp + local LAST_OUT=/tmp/tmp.uPTF4rbeQs ++ mktemp + local LAST_ERR=/tmp/tmp.Kz36ORluo6 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl logs version-service-5cd4db57-jgs4f + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.uPTF4rbeQs + cat /tmp/tmp.Kz36ORluo6 + rm /tmp/tmp.uPTF4rbeQs /tmp/tmp.Kz36ORluo6 + return 0 + diff -u /tmp/tmp.09n2hxddFJ/enabled_telemetry.version-service.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/enabled_telemetry.version-service.log.json + [[ enabled == \d\i\s\a\b\l\e\d ]] + diff -u /tmp/tmp.09n2hxddFJ/enabled_telemetry.version-service-cr.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/enabled_telemetry.version-service-cr.log.json + kubectl_bin delete perconapgcluster/minimal ++ mktemp + local LAST_OUT=/tmp/tmp.8gUB4Nmxrz ++ mktemp + local LAST_ERR=/tmp/tmp.6LKaKj88Ao + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete perconapgcluster/minimal + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.8gUB4Nmxrz perconapgcluster.pg.percona.com "minimal" deleted + cat /tmp/tmp.6LKaKj88Ao + rm /tmp/tmp.8gUB4Nmxrz /tmp/tmp.6LKaKj88Ao + return 0 ++ kubectl_bin -n version-service-20560 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cSMkdJILHt +++ mktemp ++ local LAST_ERR=/tmp/tmp.ncUDCKVqJj ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.cSMkdJILHt ++ cat /tmp/tmp.ncUDCKVqJj ++ rm /tmp/tmp.cSMkdJILHt /tmp/tmp.ncUDCKVqJj ++ return 0 + wait_job_completion minimal-rmdata-ffzc + local job=minimal-rmdata-ffzc + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' minimal-rmdata-ffzc+ retry=0 ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-ffzc -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin -n version-service-20560 delete pod --selector=run=version-service ++ mktemp + local LAST_OUT=/tmp/tmp.XkxrbxwBEx ++ mktemp + local LAST_ERR=/tmp/tmp.V9wZSmHBLh + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 delete pod --selector=run=version-service + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.XkxrbxwBEx pod "version-service-5cd4db57-jgs4f" deleted + cat /tmp/tmp.V9wZSmHBLh + rm /tmp/tmp.XkxrbxwBEx /tmp/tmp.V9wZSmHBLh + return 0 + wait_deployment version-service + local name=version-service + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace version-service + kubectl_bin -n version-service-20560 delete pod --selector=run=version-service-cr ++ mktemp + local LAST_OUT=/tmp/tmp.pN19LZN1Z0 ++ mktemp + local LAST_ERR=/tmp/tmp.qWC3ua1EDO + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 delete pod --selector=run=version-service-cr + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.pN19LZN1Z0 pod "version-service-cr-76cc4777bf-lntq9" deleted + cat /tmp/tmp.qWC3ua1EDO + rm /tmp/tmp.pN19LZN1Z0 /tmp/tmp.qWC3ua1EDO + return 0 + wait_deployment version-service-cr + local name=version-service-cr + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace version-service-cr + kubectl_bin -n version-service-20560 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"true"}}]' ++ mktemp + local LAST_OUT=/tmp/tmp.RqXKrXydWq ++ mktemp + local LAST_ERR=/tmp/tmp.ZyVKyuhOFK + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"true"}}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.RqXKrXydWq deployment.apps/postgres-operator patched + cat /tmp/tmp.ZyVKyuhOFK + rm /tmp/tmp.RqXKrXydWq /tmp/tmp.ZyVKyuhOFK + return 0 + wait_deployment postgres-operator + local name=postgres-operator + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace postgres-operator... + check_telemetry_transfer http://version-service-cr:11000 disabled disabled + local cr_vs_uri=http://version-service-cr:11000 + local cr_vs_channel=disabled + local telemetry_state=disabled + local cluster=minimal + yq w - spec.upgradeOptions.versionServiceEndpoint http://version-service-cr:11000 + yq r /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/cr-minimal.yaml + yq w - spec.upgradeOptions.apply disabled + spinup_pgcluster minimal /tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml + local cluster=minimal + local config=/tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets minimal version-service-20560 + local cluster=minimal + local namespace=version-service-20560 + yq w - stringData.some-name some-name_pass + yq w - metadata.name minimal-users-secret + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ayUIuD8U1c ++ mktemp + local LAST_ERR=/tmp/tmp.OWUmRmAhrN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ayUIuD8U1c secret/minimal-users-secret created + cat /tmp/tmp.OWUmRmAhrN + rm /tmp/tmp.ayUIuD8U1c /tmp/tmp.OWUmRmAhrN + return 0 + yq w - metadata.name minimal-pmm-secret + kubectl_bin -n version-service-20560 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.Q2H1Um5db9 ++ mktemp + local LAST_ERR=/tmp/tmp.0eHbhMVXvD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Q2H1Um5db9 secret/minimal-pmm-secret configured + cat /tmp/tmp.0eHbhMVXvD + rm /tmp/tmp.Q2H1Um5db9 /tmp/tmp.0eHbhMVXvD + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.jV2KRaa5Ni ++ mktemp + local LAST_ERR=/tmp/tmp.P3y525WYx3 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.jV2KRaa5Ni deployment.apps/pg-client unchanged + cat /tmp/tmp.P3y525WYx3 + rm /tmp/tmp.jV2KRaa5Ni /tmp/tmp.P3y525WYx3 + return 0 + apply_cluster /tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml minimal false false false false false false false false + local path=/tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml + local name=minimal + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w - metadata.labels.crunchy-pgha-scope minimal + yq w - metadata.labels.name minimal + yq w - metadata.labels.pg-cluster minimal + yq w - spec.name minimal + yq w - spec.keepBackups false + yq w - spec.keepData false + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - metadata.name minimal + yq w - spec.user some-name + yq w - spec.database minimal + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.clustername minimal + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - metadata.labels.deployment-name minimal + yq w /tmp/tmp.09n2hxddFJ/cr.minimal.apply.disabled.yaml metadata.annotations.current-primary minimal + yq w - spec.secretsName minimal-users-secret + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.ZeAp4dGTF9 ++ mktemp + local LAST_ERR=/tmp/tmp.FfsW32KQMh + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ZeAp4dGTF9 perconapgcluster.pg.percona.com/minimal created + cat /tmp/tmp.FfsW32KQMh + rm /tmp/tmp.ZeAp4dGTF9 /tmp/tmp.FfsW32KQMh + return 0 + wait_deployment minimal-backrest-shared-repo + local name=minimal-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace minimal-backrest-shared-repo.... + wait_deployment minimal + local name=minimal + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace minimal........ + [[ false == \g\c\s\+* ]] + wait_job_completion minimal-stanza-create false + local job=minimal-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' minimal-stanza-create+ retry=0 ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-minimal + local job=backrest-backup-minimal + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' backrest-backup-minimal+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency minimal + cluster_name=minimal + wait_cluster_status minimal 'pgcluster Initialized' + local cluster_name=minimal + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rAnmdxvWua +++ mktemp ++ local LAST_ERR=/tmp/tmp.v5GrX2En3I ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.rAnmdxvWua ++ cat /tmp/tmp.v5GrX2En3I ++ rm /tmp/tmp.rAnmdxvWua /tmp/tmp.v5GrX2En3I ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.izepIn16eg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZG0VKpv0yz +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.izepIn16eg +++ cat /tmp/tmp.ZG0VKpv0yz +++ rm /tmp/tmp.izepIn16eg /tmp/tmp.ZG0VKpv0yz +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZYyQip9aZp ++++ mktemp +++ local LAST_ERR=/tmp/tmp.dIRRZ3iqIx +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZYyQip9aZp +++ cat /tmp/tmp.dIRRZ3iqIx +++ rm /tmp/tmp.ZYyQip9aZp /tmp/tmp.dIRRZ3iqIx +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.PvH16wErD5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.UcbQ1Fkefu +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.PvH16wErD5 +++ cat /tmp/tmp.UcbQ1Fkefu +++ rm /tmp/tmp.PvH16wErD5 /tmp/tmp.UcbQ1Fkefu +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.uc89LeE92l ++++ mktemp +++ local LAST_ERR=/tmp/tmp.dsVMB4lh7a +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.uc89LeE92l +++ cat /tmp/tmp.dsVMB4lh7a +++ rm /tmp/tmp.uc89LeE92l /tmp/tmp.dsVMB4lh7a +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.YpuDZ7vHRD ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yqe1ObtUSs +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.YpuDZ7vHRD +++ cat /tmp/tmp.yqe1ObtUSs +++ rm /tmp/tmp.YpuDZ7vHRD /tmp/tmp.yqe1ObtUSs +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.3qFFskT8gn +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.nG3ZKG7RE0 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.3qFFskT8gn ++++ cat /tmp/tmp.nG3ZKG7RE0 ++++ rm /tmp/tmp.3qFFskT8gn /tmp/tmp.nG3ZKG7RE0 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.UmJvRYN4pY ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZsL06noVIZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.UmJvRYN4pY +++ cat /tmp/tmp.ZsL06noVIZ +++ rm /tmp/tmp.UmJvRYN4pY /tmp/tmp.ZsL06noVIZ +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@minimal.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qNdsqPU7Ot +++ mktemp ++ local LAST_ERR=/tmp/tmp.LUxUSKvtQ5 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.qNdsqPU7Ot ++ cat /tmp/tmp.LUxUSKvtQ5 ++ rm /tmp/tmp.qNdsqPU7Ot /tmp/tmp.LUxUSKvtQ5 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.eZkQb2bt15 ++ mktemp + local LAST_ERR=/tmp/tmp.1UQ18qHITP + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.eZkQb2bt15 + cat /tmp/tmp.1UQ18qHITP + rm /tmp/tmp.eZkQb2bt15 /tmp/tmp.1UQ18qHITP + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.evmN78h3Hc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.qOJbhSi6U1 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.evmN78h3Hc +++ cat /tmp/tmp.qOJbhSi6U1 +++ rm /tmp/tmp.evmN78h3Hc /tmp/tmp.qOJbhSi6U1 +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VpGqyIhTX9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.MCZX99cgd2 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.VpGqyIhTX9 +++ cat /tmp/tmp.MCZX99cgd2 +++ rm /tmp/tmp.VpGqyIhTX9 /tmp/tmp.MCZX99cgd2 +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Wq5CpB995o ++++ mktemp +++ local LAST_ERR=/tmp/tmp.arHnCM64u9 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Wq5CpB995o +++ cat /tmp/tmp.arHnCM64u9 +++ rm /tmp/tmp.Wq5CpB995o /tmp/tmp.arHnCM64u9 +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.USLrrvYCYl +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.HC5nfTitq3 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.USLrrvYCYl ++++ cat /tmp/tmp.HC5nfTitq3 ++++ rm /tmp/tmp.USLrrvYCYl /tmp/tmp.HC5nfTitq3 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.sBm396SJOc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.g3fKhT5fiN +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.sBm396SJOc +++ cat /tmp/tmp.g3fKhT5fiN +++ rm /tmp/tmp.sBm396SJOc /tmp/tmp.g3fKhT5fiN +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@minimal.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kMXOVXQr8T +++ mktemp ++ local LAST_ERR=/tmp/tmp.MQ5eEhYUDf ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.kMXOVXQr8T ++ cat /tmp/tmp.MQ5eEhYUDf ++ rm /tmp/tmp.kMXOVXQr8T /tmp/tmp.MQ5eEhYUDf ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.MPZ9uOTN54 ++ mktemp + local LAST_ERR=/tmp/tmp.AKK1HQc2yh + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.MPZ9uOTN54 + cat /tmp/tmp.AKK1HQc2yh + rm /tmp/tmp.MPZ9uOTN54 /tmp/tmp.AKK1HQc2yh + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SmxbpJmmmi ++++ mktemp +++ local LAST_ERR=/tmp/tmp.JE9jnoHv46 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.SmxbpJmmmi +++ cat /tmp/tmp.JE9jnoHv46 +++ rm /tmp/tmp.SmxbpJmmmi /tmp/tmp.JE9jnoHv46 +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.zUIjSm4TCN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.YI7w55w59Y +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.zUIjSm4TCN +++ cat /tmp/tmp.YI7w55w59Y +++ rm /tmp/tmp.zUIjSm4TCN /tmp/tmp.YI7w55w59Y +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.GwdaIypqWH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.iRemkwhdkZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.GwdaIypqWH +++ cat /tmp/tmp.iRemkwhdkZ +++ rm /tmp/tmp.GwdaIypqWH /tmp/tmp.iRemkwhdkZ +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.aU26V7nMOH +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.fEdE4VzM2S ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.aU26V7nMOH ++++ cat /tmp/tmp.fEdE4VzM2S ++++ rm /tmp/tmp.aU26V7nMOH /tmp/tmp.fEdE4VzM2S ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.j1A5X97nzd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CL86unORmO +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.j1A5X97nzd +++ cat /tmp/tmp.CL86unORmO +++ rm /tmp/tmp.j1A5X97nzd /tmp/tmp.CL86unORmO +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@minimal.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ScMr23XyfK +++ mktemp ++ local LAST_ERR=/tmp/tmp.VnJwxEhLFX ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.ScMr23XyfK ++ cat /tmp/tmp.VnJwxEhLFX ++ rm /tmp/tmp.ScMr23XyfK /tmp/tmp.VnJwxEhLFX ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.6RqrPZ00F3 ++ mktemp + local LAST_ERR=/tmp/tmp.tGlKCRVv5x + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.6RqrPZ00F3 + cat /tmp/tmp.tGlKCRVv5x + rm /tmp/tmp.6RqrPZ00F3 /tmp/tmp.tGlKCRVv5x + return 0 + sleep 10 + grep -E 'server request payload|unary call' + grep -Eo '\{.*\}' + jq 'del(."grpc.request.content".msg.customResourceUid)' + jq 'del(."grpc.time_ms")' ++ kubectl get pods --selector=run=version-service-cr -o 'jsonpath={.items[0].metadata.name}' + jq 'del(."grpc.start_time")' + kubectl_bin logs version-service-cr-76cc4777bf-tz7n7 ++ mktemp + local LAST_OUT=/tmp/tmp.PLbmSSfCLL ++ mktemp + local LAST_ERR=/tmp/tmp.6R9u1ZL1Mz + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl logs version-service-cr-76cc4777bf-tz7n7 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PLbmSSfCLL + cat /tmp/tmp.6R9u1ZL1Mz + rm /tmp/tmp.PLbmSSfCLL /tmp/tmp.6R9u1ZL1Mz + return 0 + grep -E 'server request payload|unary call' + jq 'del(."grpc.request.content".msg.customResourceUid)' + grep -Eo '\{.*\}' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.start_time")' ++ kubectl get pods --selector=run=version-service -o 'jsonpath={.items[0].metadata.name}' + kubectl_bin logs version-service-5cd4db57-z5hqz ++ mktemp + local LAST_OUT=/tmp/tmp.RpwgO3pmNd ++ mktemp + local LAST_ERR=/tmp/tmp.pniIIROqYW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl logs version-service-5cd4db57-z5hqz + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.RpwgO3pmNd + cat /tmp/tmp.pniIIROqYW + rm /tmp/tmp.RpwgO3pmNd /tmp/tmp.pniIIROqYW + return 0 + diff -u /tmp/tmp.09n2hxddFJ/disabled_telemetry.version-service.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/disabled_telemetry.version-service.log.json + [[ disabled == \d\i\s\a\b\l\e\d ]] + [[ disabled == \r\e\c\o\m\m\e\n\d\e\d ]] + diff -u /tmp/tmp.09n2hxddFJ/disabled_telemetry.version-service-cr.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/disabled_telemetry.version-service-cr.log.json + kubectl_bin delete perconapgcluster/minimal ++ mktemp + local LAST_OUT=/tmp/tmp.QM4V9CMVPU ++ mktemp + local LAST_ERR=/tmp/tmp.LMGbVfiDV4 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete perconapgcluster/minimal + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.QM4V9CMVPU perconapgcluster.pg.percona.com "minimal" deleted + cat /tmp/tmp.LMGbVfiDV4 + rm /tmp/tmp.QM4V9CMVPU /tmp/tmp.LMGbVfiDV4 + return 0 ++ kubectl_bin -n version-service-20560 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y3mHAiDA6v +++ mktemp ++ local LAST_ERR=/tmp/tmp.ERrNf8sD64 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.y3mHAiDA6v ++ cat /tmp/tmp.ERrNf8sD64 ++ rm /tmp/tmp.y3mHAiDA6v /tmp/tmp.ERrNf8sD64 ++ return 0 + wait_job_completion minimal-rmdata-qmdf + local job=minimal-rmdata-qmdf + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' minimal-rmdata-qmdf+ retry=0 ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-qmdf -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + check_telemetry_transfer http://version-service-cr:11000 recommended disabled + local cr_vs_uri=http://version-service-cr:11000 + local cr_vs_channel=recommended + local telemetry_state=disabled + local cluster=minimal + yq r /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/cr-minimal.yaml + yq w - spec.upgradeOptions.versionServiceEndpoint http://version-service-cr:11000 + yq w - spec.upgradeOptions.apply recommended + spinup_pgcluster minimal /tmp/tmp.09n2hxddFJ/cr.minimal.apply.recommended.yaml + local cluster=minimal + local config=/tmp/tmp.09n2hxddFJ/cr.minimal.apply.recommended.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets minimal version-service-20560 + local cluster=minimal + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + kubectl_bin -n version-service-20560 apply -f - + yq w - stringData.some-name some-name_pass + yq w - metadata.name minimal-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.MLEp8IL1lC ++ mktemp + local LAST_ERR=/tmp/tmp.YVEZ4MK4l1 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.MLEp8IL1lC secret/minimal-users-secret created + cat /tmp/tmp.YVEZ4MK4l1 + rm /tmp/tmp.MLEp8IL1lC /tmp/tmp.YVEZ4MK4l1 + return 0 + yq w - metadata.name minimal-pmm-secret + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.XbL0YUdgqA ++ mktemp + local LAST_ERR=/tmp/tmp.VmDq56rcXH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.XbL0YUdgqA secret/minimal-pmm-secret configured + cat /tmp/tmp.VmDq56rcXH + rm /tmp/tmp.XbL0YUdgqA /tmp/tmp.VmDq56rcXH + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.W1ui31FONM ++ mktemp + local LAST_ERR=/tmp/tmp.p4PK2rcEae + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.W1ui31FONM deployment.apps/pg-client unchanged + cat /tmp/tmp.p4PK2rcEae + rm /tmp/tmp.W1ui31FONM /tmp/tmp.p4PK2rcEae + return 0 + apply_cluster /tmp/tmp.09n2hxddFJ/cr.minimal.apply.recommended.yaml minimal false false false false false false false false + local path=/tmp/tmp.09n2hxddFJ/cr.minimal.apply.recommended.yaml + local name=minimal + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w - metadata.labels.crunchy-pgha-scope minimal + yq w - metadata.labels.deployment-name minimal + yq w - metadata.labels.pg-cluster minimal + yq w - metadata.name minimal + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.database minimal + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.name minimal + yq w - spec.keepData false + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.clustername minimal + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.secretsName minimal-users-secret + yq w - spec.user some-name + yq w - spec.keepBackups false + yq w /tmp/tmp.09n2hxddFJ/cr.minimal.apply.recommended.yaml metadata.annotations.current-primary minimal + yq w - metadata.labels.name minimal + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.46W2Sp2HZE ++ mktemp + local LAST_ERR=/tmp/tmp.NRfgoAvpg0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.46W2Sp2HZE perconapgcluster.pg.percona.com/minimal created + cat /tmp/tmp.NRfgoAvpg0 + rm /tmp/tmp.46W2Sp2HZE /tmp/tmp.NRfgoAvpg0 + return 0 + wait_deployment minimal-backrest-shared-repo + local name=minimal-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace minimal-backrest-shared-repo. + wait_deployment minimal + local name=minimal + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace minimal..... + [[ false == \g\c\s\+* ]] + wait_job_completion minimal-stanza-create false + local job=minimal-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' minimal-stanza-create+ retry=0 ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-minimal + local job=backrest-backup-minimal + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' backrest-backup-minimal+ retry=0 ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-minimal -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency minimal + cluster_name=minimal + wait_cluster_status minimal 'pgcluster Initialized' + local cluster_name=minimal + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f3HuS5VKgG +++ mktemp ++ local LAST_ERR=/tmp/tmp.sPNsGkSH5j ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.f3HuS5VKgG ++ cat /tmp/tmp.sPNsGkSH5j ++ rm /tmp/tmp.f3HuS5VKgG /tmp/tmp.sPNsGkSH5j ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.w8qNbD1Q5c ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5LcTJik1kg +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.w8qNbD1Q5c +++ cat /tmp/tmp.5LcTJik1kg +++ rm /tmp/tmp.w8qNbD1Q5c /tmp/tmp.5LcTJik1kg +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.7RpeS2iN03 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Uvlw9enDm9 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.7RpeS2iN03 +++ cat /tmp/tmp.Uvlw9enDm9 +++ rm /tmp/tmp.7RpeS2iN03 /tmp/tmp.Uvlw9enDm9 +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.RLyxP96giw ++++ mktemp +++ local LAST_ERR=/tmp/tmp.lFTP7eC3uM +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.RLyxP96giw +++ cat /tmp/tmp.lFTP7eC3uM +++ rm /tmp/tmp.RLyxP96giw /tmp/tmp.lFTP7eC3uM +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ciyRmr5MQq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Ez7d1ZSvQw +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ciyRmr5MQq +++ cat /tmp/tmp.Ez7d1ZSvQw +++ rm /tmp/tmp.ciyRmr5MQq /tmp/tmp.Ez7d1ZSvQw +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.sTr0vJH3g1 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.yFgRpSbB2C ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.sTr0vJH3g1 ++++ cat /tmp/tmp.yFgRpSbB2C ++++ rm /tmp/tmp.sTr0vJH3g1 /tmp/tmp.yFgRpSbB2C ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Ir0HVpQYdm ++++ mktemp +++ local LAST_ERR=/tmp/tmp.K19uj82AVW +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Ir0HVpQYdm +++ cat /tmp/tmp.K19uj82AVW +++ rm /tmp/tmp.Ir0HVpQYdm /tmp/tmp.K19uj82AVW +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@minimal.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hrpnDeq8V9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.wzsJfAFjMI ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.hrpnDeq8V9 ++ cat /tmp/tmp.wzsJfAFjMI ++ rm /tmp/tmp.hrpnDeq8V9 /tmp/tmp.wzsJfAFjMI ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.yDe7DkzUG6 ++ mktemp + local LAST_ERR=/tmp/tmp.r26yLEwVyi + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yDe7DkzUG6 + cat /tmp/tmp.r26yLEwVyi + rm /tmp/tmp.yDe7DkzUG6 /tmp/tmp.r26yLEwVyi + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.BzrvTkKRWz ++++ mktemp +++ local LAST_ERR=/tmp/tmp.q7j3VW7biV +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.BzrvTkKRWz +++ cat /tmp/tmp.q7j3VW7biV +++ rm /tmp/tmp.BzrvTkKRWz /tmp/tmp.q7j3VW7biV +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WpR7ssd7vq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ShdhFPanbN +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.WpR7ssd7vq +++ cat /tmp/tmp.ShdhFPanbN +++ rm /tmp/tmp.WpR7ssd7vq /tmp/tmp.ShdhFPanbN +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6Q4YqG2iXg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LHpKFlg9V0 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.6Q4YqG2iXg +++ cat /tmp/tmp.LHpKFlg9V0 +++ rm /tmp/tmp.6Q4YqG2iXg /tmp/tmp.LHpKFlg9V0 +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.g3AB0AlSzF +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.FVKwnaTzCw ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.g3AB0AlSzF ++++ cat /tmp/tmp.FVKwnaTzCw ++++ rm /tmp/tmp.g3AB0AlSzF /tmp/tmp.FVKwnaTzCw ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.nVvw3BHAvX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.3inhaKscy5 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.nVvw3BHAvX +++ cat /tmp/tmp.3inhaKscy5 +++ rm /tmp/tmp.nVvw3BHAvX /tmp/tmp.3inhaKscy5 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@minimal.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.so9bnI5jLm +++ mktemp ++ local LAST_ERR=/tmp/tmp.YO8mwS7Lhg ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.so9bnI5jLm ++ cat /tmp/tmp.YO8mwS7Lhg ++ rm /tmp/tmp.so9bnI5jLm /tmp/tmp.YO8mwS7Lhg ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.h2IhMwHxC2 ++ mktemp + local LAST_ERR=/tmp/tmp.tW7EUiDFgg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.h2IhMwHxC2 + cat /tmp/tmp.tW7EUiDFgg + rm /tmp/tmp.h2IhMwHxC2 /tmp/tmp.tW7EUiDFgg + return 0 ++ get_psql_user_pass postgres minimal ++ local user=postgres ++ local cluster=minimal ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DYp0CeEXSI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.4iYqDb9iZk +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.DYp0CeEXSI +++ cat /tmp/tmp.4iYqDb9iZk +++ rm /tmp/tmp.DYp0CeEXSI /tmp/tmp.4iYqDb9iZk +++ return 0 ++ [[ xminimal == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1AMVEGaH5L ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VZgPmXolwp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/minimal -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.1AMVEGaH5L +++ cat /tmp/tmp.VZgPmXolwp +++ rm /tmp/tmp.1AMVEGaH5L /tmp/tmp.VZgPmXolwp +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ficyLUHblD ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TelGDKIL8Y +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ficyLUHblD +++ cat /tmp/tmp.TelGDKIL8Y +++ rm /tmp/tmp.ficyLUHblD /tmp/tmp.TelGDKIL8Y +++ return 0 ++ [[ xminimal-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.jzGsmCh9Hb +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.uLvg9UqMe5 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/minimal -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.jzGsmCh9Hb ++++ cat /tmp/tmp.uLvg9UqMe5 ++++ rm /tmp/tmp.jzGsmCh9Hb /tmp/tmp.uLvg9UqMe5 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZpJhsAFs25 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.y2xUTtyZrk +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/minimal-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZpJhsAFs25 +++ cat /tmp/tmp.y2xUTtyZrk +++ rm /tmp/tmp.ZpJhsAFs25 /tmp/tmp.y2xUTtyZrk +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@minimal.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@minimal.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JEpWeNvir4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SVcGjqjN20 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.JEpWeNvir4 ++ cat /tmp/tmp.SVcGjqjN20 ++ rm /tmp/tmp.JEpWeNvir4 /tmp/tmp.SVcGjqjN20 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.pcE8tN9Wet ++ mktemp + local LAST_ERR=/tmp/tmp.cP0Qbh7ium + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@minimal.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.pcE8tN9Wet + cat /tmp/tmp.cP0Qbh7ium + rm /tmp/tmp.pcE8tN9Wet /tmp/tmp.cP0Qbh7ium + return 0 + sleep 10 + grep -E 'server request payload|unary call' + grep -Eo '\{.*\}' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.request.content".msg.customResourceUid)' + jq 'del(."grpc.start_time")' ++ kubectl get pods --selector=run=version-service-cr -o 'jsonpath={.items[0].metadata.name}' + kubectl_bin logs version-service-cr-76cc4777bf-tz7n7 ++ mktemp + local LAST_OUT=/tmp/tmp.bTMStAl23v ++ mktemp + local LAST_ERR=/tmp/tmp.pEArvscY1s + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl logs version-service-cr-76cc4777bf-tz7n7 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.bTMStAl23v + cat /tmp/tmp.pEArvscY1s + rm /tmp/tmp.bTMStAl23v /tmp/tmp.pEArvscY1s + return 0 + grep -E 'server request payload|unary call' + jq 'del(."grpc.time_ms")' + jq 'del(."grpc.start_time")' + grep -Eo '\{.*\}' + jq 'del(."grpc.request.content".msg.customResourceUid)' ++ kubectl get pods --selector=run=version-service -o 'jsonpath={.items[0].metadata.name}' + kubectl_bin logs version-service-5cd4db57-z5hqz ++ mktemp + local LAST_OUT=/tmp/tmp.gcakNHPMoc ++ mktemp + local LAST_ERR=/tmp/tmp.cjcvyoKMfe + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl logs version-service-5cd4db57-z5hqz + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gcakNHPMoc + cat /tmp/tmp.cjcvyoKMfe + rm /tmp/tmp.gcakNHPMoc /tmp/tmp.cjcvyoKMfe + return 0 + diff -u /tmp/tmp.09n2hxddFJ/disabled_telemetry.version-service.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/disabled_telemetry.version-service.log.json + [[ disabled == \d\i\s\a\b\l\e\d ]] + [[ recommended == \r\e\c\o\m\m\e\n\d\e\d ]] + diff -u /tmp/tmp.09n2hxddFJ/enabled_telemetry.version-service-cr.log.json /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/enabled_telemetry.version-service-cr.log.json + kubectl_bin delete perconapgcluster/minimal ++ mktemp + local LAST_OUT=/tmp/tmp.1E1MyBSbvo ++ mktemp + local LAST_ERR=/tmp/tmp.dceyUgNDpp + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete perconapgcluster/minimal + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.1E1MyBSbvo perconapgcluster.pg.percona.com "minimal" deleted + cat /tmp/tmp.dceyUgNDpp + rm /tmp/tmp.1E1MyBSbvo /tmp/tmp.dceyUgNDpp + return 0 ++ kubectl_bin -n version-service-20560 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bmzs5FzZ3s +++ mktemp ++ local LAST_ERR=/tmp/tmp.jNhRnQiicI ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get job --selector=pg-cluster=minimal,pgrmdata=true -o 'jsonpath={.items[0].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.bmzs5FzZ3s ++ cat /tmp/tmp.jNhRnQiicI ++ rm /tmp/tmp.bmzs5FzZ3s /tmp/tmp.jNhRnQiicI ++ return 0 + wait_job_completion minimal-rmdata-srhd + local job=minimal-rmdata-srhd + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' minimal-rmdata-srhd+ retry=0 ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/minimal-rmdata-srhd -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + kubectl_bin -n version-service-20560 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"false"}}]' ++ mktemp + local LAST_OUT=/tmp/tmp.H7GjVi9aur ++ mktemp + local LAST_ERR=/tmp/tmp.FqAx9XKert + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 patch deployment/postgres-operator --type json '-p=[{"op":"replace","path":"/spec/template/spec/containers/1/env/7","value":{"name":"DISABLE_TELEMETRY","value":"false"}}]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.H7GjVi9aur deployment.apps/postgres-operator patched + cat /tmp/tmp.FqAx9XKert + rm /tmp/tmp.H7GjVi9aur /tmp/tmp.FqAx9XKert + return 0 + wait_deployment postgres-operator + local name=postgres-operator + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace postgres-operator.. + desc 'Use simple recommended channel' + set +o xtrace ----------------------------------------------------------------------------------- Use simple recommended channel ----------------------------------------------------------------------------------- + cluster=cluster1 + PGO_TAG=1.5.1 + PG_VER=13 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-20560 apply -f - + yq w - metadata.name cluster1-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.ljihQfw7Ih ++ mktemp + local LAST_ERR=/tmp/tmp.Bm9fzWoi7C + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ljihQfw7Ih secret/cluster1-users-secret created + cat /tmp/tmp.Bm9fzWoi7C + rm /tmp/tmp.ljihQfw7Ih /tmp/tmp.Bm9fzWoi7C + return 0 + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.l2esy16qdD ++ mktemp + local LAST_ERR=/tmp/tmp.eDKgGM4WdN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.l2esy16qdD secret/cluster1-pmm-secret created + cat /tmp/tmp.eDKgGM4WdN + rm /tmp/tmp.l2esy16qdD /tmp/tmp.eDKgGM4WdN + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.qYdxSYMfWR ++ mktemp + local LAST_ERR=/tmp/tmp.RR7qVaQZ6S + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.qYdxSYMfWR deployment.apps/pg-client unchanged + cat /tmp/tmp.RR7qVaQZ6S + rm /tmp/tmp.qYdxSYMfWR /tmp/tmp.RR7qVaQZ6S + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.name cluster1 + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + yq w - spec.database cluster1 + yq w - spec.clustername cluster1 + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + yq w - metadata.labels.name cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + yq w - spec.name cluster1 + yq w - spec.keepData false + yq w - spec.keepBackups false + yq w - spec.user some-name + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - metadata.labels.deployment-name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.2dlW2Snh4L ++ mktemp + local LAST_ERR=/tmp/tmp.kfepC9MpBO + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.2dlW2Snh4L perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.kfepC9MpBO + rm /tmp/tmp.2dlW2Snh4L /tmp/tmp.kfepC9MpBO + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1........... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A0PPYFAw0y +++ mktemp ++ local LAST_ERR=/tmp/tmp.MmZrIdZUnW ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.A0PPYFAw0y ++ cat /tmp/tmp.MmZrIdZUnW ++ rm /tmp/tmp.A0PPYFAw0y /tmp/tmp.MmZrIdZUnW ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mXS7WcQavp ++++ mktemp +++ local LAST_ERR=/tmp/tmp.3TxjC8FTjY +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.mXS7WcQavp +++ cat /tmp/tmp.3TxjC8FTjY +++ rm /tmp/tmp.mXS7WcQavp /tmp/tmp.3TxjC8FTjY +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.KNpNBmXLPV ++++ mktemp +++ local LAST_ERR=/tmp/tmp.FP6TEQfDcn +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.KNpNBmXLPV +++ cat /tmp/tmp.FP6TEQfDcn +++ rm /tmp/tmp.KNpNBmXLPV /tmp/tmp.FP6TEQfDcn +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.5kqjIM7C69 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yT4lcWn0Ld +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.5kqjIM7C69 +++ cat /tmp/tmp.yT4lcWn0Ld +++ rm /tmp/tmp.5kqjIM7C69 /tmp/tmp.yT4lcWn0Ld +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.C1Zk5gnnJd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Wg0lWpGr3D +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.C1Zk5gnnJd +++ cat /tmp/tmp.Wg0lWpGr3D +++ rm /tmp/tmp.C1Zk5gnnJd /tmp/tmp.Wg0lWpGr3D +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.gUEmEddNUy +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.TfHq7srTAB ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.gUEmEddNUy ++++ cat /tmp/tmp.TfHq7srTAB ++++ rm /tmp/tmp.gUEmEddNUy /tmp/tmp.TfHq7srTAB ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.GRM0B60MyS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.DnIrty5tLp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.GRM0B60MyS +++ cat /tmp/tmp.DnIrty5tLp +++ rm /tmp/tmp.GRM0B60MyS /tmp/tmp.DnIrty5tLp +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MPtSi964tN +++ mktemp ++ local LAST_ERR=/tmp/tmp.FO8GJZXXLD ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.MPtSi964tN ++ cat /tmp/tmp.FO8GJZXXLD ++ rm /tmp/tmp.MPtSi964tN /tmp/tmp.FO8GJZXXLD ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.pUDex05XKc ++ mktemp + local LAST_ERR=/tmp/tmp.7Ijhd3d2MN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.pUDex05XKc + cat /tmp/tmp.7Ijhd3d2MN + rm /tmp/tmp.pUDex05XKc /tmp/tmp.7Ijhd3d2MN + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.UGFeEjAGOW ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wp5vNEJaUZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.UGFeEjAGOW +++ cat /tmp/tmp.wp5vNEJaUZ +++ rm /tmp/tmp.UGFeEjAGOW /tmp/tmp.wp5vNEJaUZ +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.vCdAI9tUiD ++++ mktemp +++ local LAST_ERR=/tmp/tmp.eLSzd3lUCS +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.vCdAI9tUiD +++ cat /tmp/tmp.eLSzd3lUCS +++ rm /tmp/tmp.vCdAI9tUiD /tmp/tmp.eLSzd3lUCS +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.oyDxgVeltv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jC3RmgZJQw +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.oyDxgVeltv +++ cat /tmp/tmp.jC3RmgZJQw +++ rm /tmp/tmp.oyDxgVeltv /tmp/tmp.jC3RmgZJQw +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.hsGQthy2aA +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.bxK1MquqMc ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.hsGQthy2aA ++++ cat /tmp/tmp.bxK1MquqMc ++++ rm /tmp/tmp.hsGQthy2aA /tmp/tmp.bxK1MquqMc ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.erFJNUIeKH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Q572xMVycM +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.erFJNUIeKH +++ cat /tmp/tmp.Q572xMVycM +++ rm /tmp/tmp.erFJNUIeKH /tmp/tmp.Q572xMVycM +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l2xOv8zHDS +++ mktemp ++ local LAST_ERR=/tmp/tmp.wGgd2jfwbv ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.l2xOv8zHDS ++ cat /tmp/tmp.wGgd2jfwbv ++ rm /tmp/tmp.l2xOv8zHDS /tmp/tmp.wGgd2jfwbv ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.kRTZC8FdqE ++ mktemp + local LAST_ERR=/tmp/tmp.0a2j7Kvjml + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.kRTZC8FdqE + cat /tmp/tmp.0a2j7Kvjml + rm /tmp/tmp.kRTZC8FdqE /tmp/tmp.0a2j7Kvjml + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qY3NL5KdvV ++++ mktemp +++ local LAST_ERR=/tmp/tmp.7fj0fHkI3n +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qY3NL5KdvV +++ cat /tmp/tmp.7fj0fHkI3n +++ rm /tmp/tmp.qY3NL5KdvV /tmp/tmp.7fj0fHkI3n +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Rgl1TldpSC ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yZBhfx2GhM +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Rgl1TldpSC +++ cat /tmp/tmp.yZBhfx2GhM +++ rm /tmp/tmp.Rgl1TldpSC /tmp/tmp.yZBhfx2GhM +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.f7m8MipNsZ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.HIokZDUnRP +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.f7m8MipNsZ +++ cat /tmp/tmp.HIokZDUnRP +++ rm /tmp/tmp.f7m8MipNsZ /tmp/tmp.HIokZDUnRP +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.ZD5spI4USu +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Wv4iIDmDhX ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.ZD5spI4USu ++++ cat /tmp/tmp.Wv4iIDmDhX ++++ rm /tmp/tmp.ZD5spI4USu /tmp/tmp.Wv4iIDmDhX ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Vy1V5jAJ03 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.kLme4vjbir +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Vy1V5jAJ03 +++ cat /tmp/tmp.kLme4vjbir +++ rm /tmp/tmp.Vy1V5jAJ03 /tmp/tmp.kLme4vjbir +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rr0JOfkuhp +++ mktemp ++ local LAST_ERR=/tmp/tmp.oYM95Sqa30 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.rr0JOfkuhp ++ cat /tmp/tmp.oYM95Sqa30 ++ rm /tmp/tmp.rr0JOfkuhp /tmp/tmp.oYM95Sqa30 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.41lBEm556H ++ mktemp + local LAST_ERR=/tmp/tmp.FvdR2edi10 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.41lBEm556H + cat /tmp/tmp.FvdR2edi10 + rm /tmp/tmp.41lBEm556H /tmp/tmp.FvdR2edi10 + return 0 + sleep 10 + apply_version cluster1 9.9.9 recommended 13 http://version-service:11000 0 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=recommended + local db_version=13 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=0 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.Os2Aw3R4ih ++ mktemp + local LAST_ERR=/tmp/tmp.3UeRRHCggZ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Os2Aw3R4ih perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.3UeRRHCggZ + rm /tmp/tmp.Os2Aw3R4ih /tmp/tmp.3UeRRHCggZ + return 0 + sleep 10 + wait_smart_update cluster1 0 + local cluster=cluster1 + local target=0 + set +o xtrace ......................................................................... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.cf9PJ8j7PA ++ mktemp + local LAST_ERR=/tmp/tmp.XCS4nN7KxJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.cf9PJ8j7PA job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.XCS4nN7KxJ + rm /tmp/tmp.cf9PJ8j7PA /tmp/tmp.XCS4nN7KxJ + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - spec.namespace version-service-20560 + yq w - spec.parameters.job-name fulldata-cluster1 + kubectl_bin apply -f - + yq w --style=single -- - spec.parameters.backrest-opts --type=full + yq w - spec.parameters.pg-cluster cluster1 ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.YBcYnWsymd ++ local LAST_OUT=/tmp/tmp.TNUW8X5oFY +++ mktemp ++ local LAST_ERR=/tmp/tmp.r924k2oPV5 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ mktemp + local LAST_ERR=/tmp/tmp.9RPgtPOnNF + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.TNUW8X5oFY ++ cat /tmp/tmp.r924k2oPV5 ++ rm /tmp/tmp.TNUW8X5oFY /tmp/tmp.r924k2oPV5 ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-57f668d478-qsc7n + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.YBcYnWsymd pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.9RPgtPOnNF + rm /tmp/tmp.YBcYnWsymd /tmp/tmp.9RPgtPOnNF + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 recommended 13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=recommended + local db_version=13 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.7keIBe5Rcv ++ mktemp + local LAST_ERR=/tmp/tmp.55EpUawOrw + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.7keIBe5Rcv + cat /tmp/tmp.55EpUawOrw + rm /tmp/tmp.7keIBe5Rcv /tmp/tmp.55EpUawOrw + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Use major version recommended channel' + set +o xtrace ----------------------------------------------------------------------------------- Use major version recommended channel ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + kubectl_bin -n version-service-20560 apply -f - + yq w - stringData.some-name some-name_pass ++ mktemp + local LAST_OUT=/tmp/tmp.p3ZztWzgdv ++ mktemp + local LAST_ERR=/tmp/tmp.W6gzloBf3U + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.p3ZztWzgdv secret/cluster1-users-secret created + cat /tmp/tmp.W6gzloBf3U + rm /tmp/tmp.p3ZztWzgdv /tmp/tmp.W6gzloBf3U + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.p4s43eM9q7 ++ mktemp + local LAST_ERR=/tmp/tmp.eZtDRcqgwT + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.p4s43eM9q7 secret/cluster1-pmm-secret configured + cat /tmp/tmp.eZtDRcqgwT + rm /tmp/tmp.p4s43eM9q7 /tmp/tmp.eZtDRcqgwT + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.s7ePEkDJ9W ++ mktemp + local LAST_ERR=/tmp/tmp.vbmeTZ6smJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.s7ePEkDJ9W deployment.apps/pg-client unchanged + cat /tmp/tmp.vbmeTZ6smJ + rm /tmp/tmp.s7ePEkDJ9W /tmp/tmp.vbmeTZ6smJ + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - spec.database cluster1 + yq w - spec.name cluster1 + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + yq w - metadata.name cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - spec.clustername cluster1 + yq w - spec.keepData false + yq w - spec.keepBackups false + yq w - spec.user some-name + yq w - metadata.labels.name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.yDqJ2oOpdF ++ mktemp + local LAST_ERR=/tmp/tmp.jSVJ01HMau + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yDqJ2oOpdF perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.jSVJ01HMau + rm /tmp/tmp.yDqJ2oOpdF /tmp/tmp.jSVJ01HMau + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1............ + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H2084RnyIR +++ mktemp ++ local LAST_ERR=/tmp/tmp.J4fuMXawBj ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.H2084RnyIR ++ cat /tmp/tmp.J4fuMXawBj ++ rm /tmp/tmp.H2084RnyIR /tmp/tmp.J4fuMXawBj ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.GWBk8citGf ++++ mktemp +++ local LAST_ERR=/tmp/tmp.YBz6lU33pk +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.GWBk8citGf +++ cat /tmp/tmp.YBz6lU33pk +++ rm /tmp/tmp.GWBk8citGf /tmp/tmp.YBz6lU33pk +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.5CWMjnqRK6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Yf0x0NTuSc +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.5CWMjnqRK6 +++ cat /tmp/tmp.Yf0x0NTuSc +++ rm /tmp/tmp.5CWMjnqRK6 /tmp/tmp.Yf0x0NTuSc +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6a4L10L2E7 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Bd7XPNKMv8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.6a4L10L2E7 +++ cat /tmp/tmp.Bd7XPNKMv8 +++ rm /tmp/tmp.6a4L10L2E7 /tmp/tmp.Bd7XPNKMv8 +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.k0Am2B47QK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.xUX1Z8gdUs +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.k0Am2B47QK +++ cat /tmp/tmp.xUX1Z8gdUs +++ rm /tmp/tmp.k0Am2B47QK /tmp/tmp.xUX1Z8gdUs +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.tiSyh9021N +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.0JUUiy8uR9 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.tiSyh9021N ++++ cat /tmp/tmp.0JUUiy8uR9 ++++ rm /tmp/tmp.tiSyh9021N /tmp/tmp.0JUUiy8uR9 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.oQOWAkqRbl ++++ mktemp +++ local LAST_ERR=/tmp/tmp.usyzWUt2gF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.oQOWAkqRbl +++ cat /tmp/tmp.usyzWUt2gF +++ rm /tmp/tmp.oQOWAkqRbl /tmp/tmp.usyzWUt2gF +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uINrczC2I5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cdvSCiwqzN ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.uINrczC2I5 ++ cat /tmp/tmp.cdvSCiwqzN ++ rm /tmp/tmp.uINrczC2I5 /tmp/tmp.cdvSCiwqzN ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.HE2VNjcBES ++ mktemp + local LAST_ERR=/tmp/tmp.iXt6bfTcoH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.HE2VNjcBES + cat /tmp/tmp.iXt6bfTcoH + rm /tmp/tmp.HE2VNjcBES /tmp/tmp.iXt6bfTcoH + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.QpEo2iZmEQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.liXMqUF8O2 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.QpEo2iZmEQ +++ cat /tmp/tmp.liXMqUF8O2 +++ rm /tmp/tmp.QpEo2iZmEQ /tmp/tmp.liXMqUF8O2 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.BtdxZnlZAA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jdsVxbQBaW +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.BtdxZnlZAA +++ cat /tmp/tmp.jdsVxbQBaW +++ rm /tmp/tmp.BtdxZnlZAA /tmp/tmp.jdsVxbQBaW +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.FFQMcjJM8Q ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NHHIaUyIUp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.FFQMcjJM8Q +++ cat /tmp/tmp.NHHIaUyIUp +++ rm /tmp/tmp.FFQMcjJM8Q /tmp/tmp.NHHIaUyIUp +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.epjumMH29N +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.hJxiYytFgn ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.epjumMH29N ++++ cat /tmp/tmp.hJxiYytFgn ++++ rm /tmp/tmp.epjumMH29N /tmp/tmp.hJxiYytFgn ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qMDztP1hZA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.a7s7SlWjkr +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qMDztP1hZA +++ cat /tmp/tmp.a7s7SlWjkr +++ rm /tmp/tmp.qMDztP1hZA /tmp/tmp.a7s7SlWjkr +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tGzd3mV4LJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.FAheRHcA67 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.tGzd3mV4LJ ++ cat /tmp/tmp.FAheRHcA67 ++ rm /tmp/tmp.tGzd3mV4LJ /tmp/tmp.FAheRHcA67 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.Swg4d4zHHu ++ mktemp + local LAST_ERR=/tmp/tmp.VWEE8sJYFy + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Swg4d4zHHu + cat /tmp/tmp.VWEE8sJYFy + rm /tmp/tmp.Swg4d4zHHu /tmp/tmp.VWEE8sJYFy + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.2aXn917mPI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.axeLHOVKEn +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.2aXn917mPI +++ cat /tmp/tmp.axeLHOVKEn +++ rm /tmp/tmp.2aXn917mPI /tmp/tmp.axeLHOVKEn +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.8Fh777LckX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.8eQDydo0G8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.8Fh777LckX +++ cat /tmp/tmp.8eQDydo0G8 +++ rm /tmp/tmp.8Fh777LckX /tmp/tmp.8eQDydo0G8 +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.9D1jQpvUPe ++++ mktemp +++ local LAST_ERR=/tmp/tmp.O4n43BSUgw +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.9D1jQpvUPe +++ cat /tmp/tmp.O4n43BSUgw +++ rm /tmp/tmp.9D1jQpvUPe /tmp/tmp.O4n43BSUgw +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.ObYaJ8sDVX +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Hwz0TVKkil ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.ObYaJ8sDVX ++++ cat /tmp/tmp.Hwz0TVKkil ++++ rm /tmp/tmp.ObYaJ8sDVX /tmp/tmp.Hwz0TVKkil ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.bwPGxXGjhV ++++ mktemp +++ local LAST_ERR=/tmp/tmp.6t6o1Rx29Q +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.bwPGxXGjhV +++ cat /tmp/tmp.6t6o1Rx29Q +++ rm /tmp/tmp.bwPGxXGjhV /tmp/tmp.6t6o1Rx29Q +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WW4g9uU40R +++ mktemp ++ local LAST_ERR=/tmp/tmp.yoquBNpdFu ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.WW4g9uU40R ++ cat /tmp/tmp.yoquBNpdFu ++ rm /tmp/tmp.WW4g9uU40R /tmp/tmp.yoquBNpdFu ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.48dwjOMLBD ++ mktemp + local LAST_ERR=/tmp/tmp.9VKv1aBSrg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.48dwjOMLBD + cat /tmp/tmp.9VKv1aBSrg + rm /tmp/tmp.48dwjOMLBD /tmp/tmp.9VKv1aBSrg + return 0 + sleep 10 + apply_version cluster1 9.9.9 12-recommended 12 http://version-service:11000 1 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-recommended + local db_version=12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=1 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.xB4RIbr4I7 ++ mktemp + local LAST_ERR=/tmp/tmp.x5I5RFrTYD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xB4RIbr4I7 perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.x5I5RFrTYD + rm /tmp/tmp.xB4RIbr4I7 /tmp/tmp.x5I5RFrTYD + return 0 + sleep 10 + wait_smart_update cluster1 1 + local cluster=cluster1 + local target=1 + set +o xtrace ........................................................................... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.AhfwpoZH0P ++ mktemp + local LAST_ERR=/tmp/tmp.ZKqXBhoruc + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.AhfwpoZH0P job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.ZKqXBhoruc + rm /tmp/tmp.AhfwpoZH0P /tmp/tmp.ZKqXBhoruc + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - metadata.name fulldata-cluster1 + yq w - spec.parameters.pg-cluster cluster1 + kubectl_bin apply -f - + yq w - spec.parameters.job-name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + yq w - spec.namespace version-service-20560 ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ mktemp +++ mktemp + local LAST_OUT=/tmp/tmp.EKasa5Xnvx ++ local LAST_OUT=/tmp/tmp.7ngcPln4nR +++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.gdMpyjYjWZ + local exit_status=0 + local timeout=4 + set +o errexit ++ local LAST_ERR=/tmp/tmp.Y8zDZ5KSKN ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit ++ seq 0 2 +++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.7ngcPln4nR ++ cat /tmp/tmp.Y8zDZ5KSKN ++ rm /tmp/tmp.7ngcPln4nR /tmp/tmp.Y8zDZ5KSKN ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-5f647d77b5-zsv7l + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.EKasa5Xnvx pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.gdMpyjYjWZ + rm /tmp/tmp.EKasa5Xnvx /tmp/tmp.gdMpyjYjWZ + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 12-recommended 12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-recommended + local db_version=12 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.nd6zDVCrit ++ mktemp + local LAST_ERR=/tmp/tmp.Jr4lDsg2c5 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nd6zDVCrit + cat /tmp/tmp.Jr4lDsg2c5 + rm /tmp/tmp.nd6zDVCrit /tmp/tmp.Jr4lDsg2c5 + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-12-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ..Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Updating PG cluster with 14 from 13' + set +o xtrace ----------------------------------------------------------------------------------- Updating PG cluster with 14 from 13 ----------------------------------------------------------------------------------- + PG_VER=13 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.uGXAz2DrTH ++ mktemp + local LAST_ERR=/tmp/tmp.tGZK6Aa1w0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.uGXAz2DrTH secret/cluster1-users-secret created + cat /tmp/tmp.tGZK6Aa1w0 + rm /tmp/tmp.uGXAz2DrTH /tmp/tmp.tGZK6Aa1w0 + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.8Kw32swXxi ++ mktemp + local LAST_ERR=/tmp/tmp.kjhohxcPcM + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.8Kw32swXxi secret/cluster1-pmm-secret configured + cat /tmp/tmp.kjhohxcPcM + rm /tmp/tmp.8Kw32swXxi /tmp/tmp.kjhohxcPcM + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.nTDKpb1odM ++ mktemp + local LAST_ERR=/tmp/tmp.sB5hWvoXyS + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nTDKpb1odM deployment.apps/pg-client unchanged + cat /tmp/tmp.sB5hWvoXyS + rm /tmp/tmp.nTDKpb1odM /tmp/tmp.sB5hWvoXyS + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.name cluster1 + yq w - spec.clustername cluster1 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + yq w - spec.keepBackups false + yq w - spec.user some-name + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + yq w - spec.keepData false + yq w - spec.name cluster1 + yq w - spec.database cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.labels.deployment-name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.W9QRzI2aN8 ++ mktemp + local LAST_ERR=/tmp/tmp.gamsNYvYfx + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.W9QRzI2aN8 perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.gamsNYvYfx + rm /tmp/tmp.W9QRzI2aN8 /tmp/tmp.gamsNYvYfx + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1...... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BPutzOOnok +++ mktemp ++ local LAST_ERR=/tmp/tmp.DDywmsW2wO ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.BPutzOOnok ++ cat /tmp/tmp.DDywmsW2wO ++ rm /tmp/tmp.BPutzOOnok /tmp/tmp.DDywmsW2wO ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.HrJzJqwrFA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.2blH5htA3I +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.HrJzJqwrFA +++ cat /tmp/tmp.2blH5htA3I +++ rm /tmp/tmp.HrJzJqwrFA /tmp/tmp.2blH5htA3I +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.4MvDjIDqju ++++ mktemp +++ local LAST_ERR=/tmp/tmp.MMhvfHQT1y +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.4MvDjIDqju +++ cat /tmp/tmp.MMhvfHQT1y +++ rm /tmp/tmp.4MvDjIDqju /tmp/tmp.MMhvfHQT1y +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.a4YKp5lQ3b ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BI3LteMCU8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.a4YKp5lQ3b +++ cat /tmp/tmp.BI3LteMCU8 +++ rm /tmp/tmp.a4YKp5lQ3b /tmp/tmp.BI3LteMCU8 +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Lz6B12ZXI3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TbCmaV6kPg +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Lz6B12ZXI3 +++ cat /tmp/tmp.TbCmaV6kPg +++ rm /tmp/tmp.Lz6B12ZXI3 /tmp/tmp.TbCmaV6kPg +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.p8hPv3qv4f +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.51TGT0Ip75 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.p8hPv3qv4f ++++ cat /tmp/tmp.51TGT0Ip75 ++++ rm /tmp/tmp.p8hPv3qv4f /tmp/tmp.51TGT0Ip75 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Kh2ZOnYTYv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.07bSRRIEHR +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Kh2ZOnYTYv +++ cat /tmp/tmp.07bSRRIEHR +++ rm /tmp/tmp.Kh2ZOnYTYv /tmp/tmp.07bSRRIEHR +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TInDcBKaw8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.hHbQ2Ncc6i ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.TInDcBKaw8 ++ cat /tmp/tmp.hHbQ2Ncc6i ++ rm /tmp/tmp.TInDcBKaw8 /tmp/tmp.hHbQ2Ncc6i ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.DY0YFeR2rR ++ mktemp + local LAST_ERR=/tmp/tmp.Le7xn32w37 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.DY0YFeR2rR + cat /tmp/tmp.Le7xn32w37 + rm /tmp/tmp.DY0YFeR2rR /tmp/tmp.Le7xn32w37 + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.jwAXPQugyL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1CJHWrqOwg +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.jwAXPQugyL +++ cat /tmp/tmp.1CJHWrqOwg +++ rm /tmp/tmp.jwAXPQugyL /tmp/tmp.1CJHWrqOwg +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Rrlh4NJMIX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.hvPivBLuTL +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Rrlh4NJMIX +++ cat /tmp/tmp.hvPivBLuTL +++ rm /tmp/tmp.Rrlh4NJMIX /tmp/tmp.hvPivBLuTL +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Pjpkw0NmUE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0KsX8HWKDf +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Pjpkw0NmUE +++ cat /tmp/tmp.0KsX8HWKDf +++ rm /tmp/tmp.Pjpkw0NmUE /tmp/tmp.0KsX8HWKDf +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.eNwuv2olmw +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.pFliajvKUU ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.eNwuv2olmw ++++ cat /tmp/tmp.pFliajvKUU ++++ rm /tmp/tmp.eNwuv2olmw /tmp/tmp.pFliajvKUU ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.9T5U2d44hh ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BlrlTdRoI1 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.9T5U2d44hh +++ cat /tmp/tmp.BlrlTdRoI1 +++ rm /tmp/tmp.9T5U2d44hh /tmp/tmp.BlrlTdRoI1 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zJ2wOXdtAz +++ mktemp ++ local LAST_ERR=/tmp/tmp.vcWpAUyyNq ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.zJ2wOXdtAz ++ cat /tmp/tmp.vcWpAUyyNq ++ rm /tmp/tmp.zJ2wOXdtAz /tmp/tmp.vcWpAUyyNq ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.yNqYIBsOxH ++ mktemp + local LAST_ERR=/tmp/tmp.6SkaqFxW3N + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yNqYIBsOxH + cat /tmp/tmp.6SkaqFxW3N + rm /tmp/tmp.yNqYIBsOxH /tmp/tmp.6SkaqFxW3N + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.IXTOkjHym5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.EOq86m8OUu +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.IXTOkjHym5 +++ cat /tmp/tmp.EOq86m8OUu +++ rm /tmp/tmp.IXTOkjHym5 /tmp/tmp.EOq86m8OUu +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.atYcmPbaPV ++++ mktemp +++ local LAST_ERR=/tmp/tmp.tDRuD1B1Oz +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.atYcmPbaPV +++ cat /tmp/tmp.tDRuD1B1Oz +++ rm /tmp/tmp.atYcmPbaPV /tmp/tmp.tDRuD1B1Oz +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.HuNpFVkajb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Zd30G92KGK +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.HuNpFVkajb +++ cat /tmp/tmp.Zd30G92KGK +++ rm /tmp/tmp.HuNpFVkajb /tmp/tmp.Zd30G92KGK +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.9DuU4jssHa +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.kGm4vVVd54 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.9DuU4jssHa ++++ cat /tmp/tmp.kGm4vVVd54 ++++ rm /tmp/tmp.9DuU4jssHa /tmp/tmp.kGm4vVVd54 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.An4vyZpJcB ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vUYGriTDQU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.An4vyZpJcB +++ cat /tmp/tmp.vUYGriTDQU +++ rm /tmp/tmp.An4vyZpJcB /tmp/tmp.vUYGriTDQU +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K1tGaVPYbi +++ mktemp ++ local LAST_ERR=/tmp/tmp.WuLqVGUOf6 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.K1tGaVPYbi ++ cat /tmp/tmp.WuLqVGUOf6 ++ rm /tmp/tmp.K1tGaVPYbi /tmp/tmp.WuLqVGUOf6 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.VmgJePF2LT ++ mktemp + local LAST_ERR=/tmp/tmp.N085LlhN72 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.VmgJePF2LT + cat /tmp/tmp.N085LlhN72 + rm /tmp/tmp.VmgJePF2LT /tmp/tmp.N085LlhN72 + return 0 + sleep 10 + apply_version cluster1 9.9.9 14-recommended 1.5.1-13 http://version-service:11000 1 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=14-recommended + local db_version=1.5.1-13 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=1 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"14-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.wDiDSoAi2D ++ mktemp + local LAST_ERR=/tmp/tmp.7XyLJTRUhX + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"14-recommended"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.wDiDSoAi2D perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.7XyLJTRUhX + rm /tmp/tmp.wDiDSoAi2D /tmp/tmp.7XyLJTRUhX + return 0 + sleep 10 + wait_smart_update cluster1 1 + local cluster=cluster1 + local target=1 + set +o xtrace + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1.......... + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.zlz6CXPAyo ++ mktemp + local LAST_ERR=/tmp/tmp.1CBWTkG23H + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.zlz6CXPAyo job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.1CBWTkG23H + rm /tmp/tmp.zlz6CXPAyo /tmp/tmp.1CBWTkG23H + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.parameters.pg-cluster cluster1 + kubectl_bin apply -f - + yq w - metadata.name fulldata-cluster1 ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.EJlSqAb9iO ++ local LAST_OUT=/tmp/tmp.jYNrVbAOb9 +++ mktemp ++ mktemp ++ local LAST_ERR=/tmp/tmp.HwUgxHW1E7 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + yq w - spec.namespace version-service-20560 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + yq w - spec.parameters.job-name fulldata-cluster1 + local LAST_ERR=/tmp/tmp.yoNxB4fAuI + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.jYNrVbAOb9 ++ cat /tmp/tmp.HwUgxHW1E7 ++ rm /tmp/tmp.jYNrVbAOb9 /tmp/tmp.HwUgxHW1E7 ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-65857c588c-pg2vr + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.EJlSqAb9iO pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.yoNxB4fAuI + rm /tmp/tmp.EJlSqAb9iO /tmp/tmp.yoNxB4fAuI + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 14-recommended 1.5.1-13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=14-recommended + local db_version=1.5.1-13 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.z6GS4KDe9T ++ mktemp + local LAST_ERR=/tmp/tmp.kQXT6FM9iQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.z6GS4KDe9T + cat /tmp/tmp.kQXT6FM9iQ + rm /tmp/tmp.z6GS4KDe9T /tmp/tmp.kQXT6FM9iQ + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-14-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/1.5.1-13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster update with the latest image by version service' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster update with the latest image by version service ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq w - metadata.name cluster1-users-secret + yq w - stringData.some-name some-name_pass + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.9vpVqK79FR ++ mktemp + local LAST_ERR=/tmp/tmp.aTyC2NHYE2 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9vpVqK79FR secret/cluster1-users-secret created + cat /tmp/tmp.aTyC2NHYE2 + rm /tmp/tmp.9vpVqK79FR /tmp/tmp.aTyC2NHYE2 + return 0 + yq w - metadata.name cluster1-pmm-secret + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.OFgJN6GAvW ++ mktemp + local LAST_ERR=/tmp/tmp.BG6r9wBAAP + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.OFgJN6GAvW secret/cluster1-pmm-secret configured + cat /tmp/tmp.BG6r9wBAAP + rm /tmp/tmp.OFgJN6GAvW /tmp/tmp.BG6r9wBAAP + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.xMzuic03bj ++ mktemp + local LAST_ERR=/tmp/tmp.xVwNVNlm6x + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xMzuic03bj deployment.apps/pg-client unchanged + cat /tmp/tmp.xVwNVNlm6x + rm /tmp/tmp.xMzuic03bj /tmp/tmp.xVwNVNlm6x + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - spec.clustername cluster1 + yq w - spec.database cluster1 + yq w - spec.keepData false + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.user some-name + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + yq w - spec.keepBackups false + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + yq w - spec.name cluster1 + yq w - metadata.name cluster1 + yq w - metadata.labels.name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.Iuph4ippeP ++ mktemp + local LAST_ERR=/tmp/tmp.X87x7Y3aWI + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Iuph4ippeP perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.X87x7Y3aWI + rm /tmp/tmp.Iuph4ippeP /tmp/tmp.X87x7Y3aWI + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1...... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qJMInrHIPQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.vctJSLpaYT ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.qJMInrHIPQ ++ cat /tmp/tmp.vctJSLpaYT ++ rm /tmp/tmp.qJMInrHIPQ /tmp/tmp.vctJSLpaYT ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OcOeFX7mW1 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5vUnnnTEKm +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.OcOeFX7mW1 +++ cat /tmp/tmp.5vUnnnTEKm +++ rm /tmp/tmp.OcOeFX7mW1 /tmp/tmp.5vUnnnTEKm +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.7QqfCaVCah ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZzPszdKkvZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.7QqfCaVCah +++ cat /tmp/tmp.ZzPszdKkvZ +++ rm /tmp/tmp.7QqfCaVCah /tmp/tmp.ZzPszdKkvZ +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.8gbMEz6Cd0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Wu5fDKpkOG +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.8gbMEz6Cd0 +++ cat /tmp/tmp.Wu5fDKpkOG +++ rm /tmp/tmp.8gbMEz6Cd0 /tmp/tmp.Wu5fDKpkOG +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.4S37TN61xI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TlOs3cN0WW +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.4S37TN61xI +++ cat /tmp/tmp.TlOs3cN0WW +++ rm /tmp/tmp.4S37TN61xI /tmp/tmp.TlOs3cN0WW +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.R8KNfdg6qQ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.7ERWCbLYqn ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.R8KNfdg6qQ ++++ cat /tmp/tmp.7ERWCbLYqn ++++ rm /tmp/tmp.R8KNfdg6qQ /tmp/tmp.7ERWCbLYqn ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.4l40RWvjYx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.eqIXquUE1m +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.4l40RWvjYx +++ cat /tmp/tmp.eqIXquUE1m +++ rm /tmp/tmp.4l40RWvjYx /tmp/tmp.eqIXquUE1m +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U8tyz3kOBv +++ mktemp ++ local LAST_ERR=/tmp/tmp.wLQg0cKqnm ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.U8tyz3kOBv ++ cat /tmp/tmp.wLQg0cKqnm ++ rm /tmp/tmp.U8tyz3kOBv /tmp/tmp.wLQg0cKqnm ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.KvNpsYJWSk ++ mktemp + local LAST_ERR=/tmp/tmp.mqzMf7vgtl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.KvNpsYJWSk + cat /tmp/tmp.mqzMf7vgtl + rm /tmp/tmp.KvNpsYJWSk /tmp/tmp.mqzMf7vgtl + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.zb0pOb3YTF ++++ mktemp +++ local LAST_ERR=/tmp/tmp.YuxcLNew9g +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.zb0pOb3YTF +++ cat /tmp/tmp.YuxcLNew9g +++ rm /tmp/tmp.zb0pOb3YTF /tmp/tmp.YuxcLNew9g +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.yrCVO9V3Kt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.3Y9QB0X9yB +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.yrCVO9V3Kt +++ cat /tmp/tmp.3Y9QB0X9yB +++ rm /tmp/tmp.yrCVO9V3Kt /tmp/tmp.3Y9QB0X9yB +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.JVEuNYwTKL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.965lFwnYee +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.JVEuNYwTKL +++ cat /tmp/tmp.965lFwnYee +++ rm /tmp/tmp.JVEuNYwTKL /tmp/tmp.965lFwnYee +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.i25t4MqkJR +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.5sLwKuL4ry ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.i25t4MqkJR ++++ cat /tmp/tmp.5sLwKuL4ry ++++ rm /tmp/tmp.i25t4MqkJR /tmp/tmp.5sLwKuL4ry ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.QK0grOOKmL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.l7wSbW9Wem +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.QK0grOOKmL +++ cat /tmp/tmp.l7wSbW9Wem +++ rm /tmp/tmp.QK0grOOKmL /tmp/tmp.l7wSbW9Wem +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cmxveUizHn +++ mktemp ++ local LAST_ERR=/tmp/tmp.L5KYf6KrWc ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.cmxveUizHn ++ cat /tmp/tmp.L5KYf6KrWc ++ rm /tmp/tmp.cmxveUizHn /tmp/tmp.L5KYf6KrWc ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.g6EPGQq7KY ++ mktemp + local LAST_ERR=/tmp/tmp.MXDqiueg6H + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.g6EPGQq7KY + cat /tmp/tmp.MXDqiueg6H + rm /tmp/tmp.g6EPGQq7KY /tmp/tmp.MXDqiueg6H + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.clJAAIfJxr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5dpelfXjMU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.clJAAIfJxr +++ cat /tmp/tmp.5dpelfXjMU +++ rm /tmp/tmp.clJAAIfJxr /tmp/tmp.5dpelfXjMU +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3FlwhyA6h9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9xh3Ru5f25 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.3FlwhyA6h9 +++ cat /tmp/tmp.9xh3Ru5f25 +++ rm /tmp/tmp.3FlwhyA6h9 /tmp/tmp.9xh3Ru5f25 +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0BBjuVy2Bq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.7zk03QyDYK +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.0BBjuVy2Bq +++ cat /tmp/tmp.7zk03QyDYK +++ rm /tmp/tmp.0BBjuVy2Bq /tmp/tmp.7zk03QyDYK +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.pQjxrYcJtv +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.FQn4oC7Jbe ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.pQjxrYcJtv ++++ cat /tmp/tmp.FQn4oC7Jbe ++++ rm /tmp/tmp.pQjxrYcJtv /tmp/tmp.FQn4oC7Jbe ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.7jv9ubzvHc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.np0ONp03TI +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.7jv9ubzvHc +++ cat /tmp/tmp.np0ONp03TI +++ rm /tmp/tmp.7jv9ubzvHc /tmp/tmp.np0ONp03TI +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bdwN0x78nw +++ mktemp ++ local LAST_ERR=/tmp/tmp.th8xRjZb4Q ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.bdwN0x78nw ++ cat /tmp/tmp.th8xRjZb4Q ++ rm /tmp/tmp.bdwN0x78nw /tmp/tmp.th8xRjZb4Q ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.LmC53l7un5 ++ mktemp + local LAST_ERR=/tmp/tmp.1YdnqRlGuk + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.LmC53l7un5 + cat /tmp/tmp.1YdnqRlGuk + rm /tmp/tmp.LmC53l7un5 /tmp/tmp.1YdnqRlGuk + return 0 + sleep 10 + apply_version cluster1 9.9.9 latest 12 http://version-service:11000 2 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=2 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.0zLS0Afw4C ++ mktemp + local LAST_ERR=/tmp/tmp.TnLf5MZwmm + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.0zLS0Afw4C perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.TnLf5MZwmm + rm /tmp/tmp.0zLS0Afw4C /tmp/tmp.TnLf5MZwmm + return 0 + sleep 10 + wait_smart_update cluster1 2 + local cluster=cluster1 + local target=2 + set +o xtrace ................................................. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.D0l3UkxBBm ++ mktemp + local LAST_ERR=/tmp/tmp.ycNbYf1tG9 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.D0l3UkxBBm job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.ycNbYf1tG9 + rm /tmp/tmp.D0l3UkxBBm /tmp/tmp.ycNbYf1tG9 + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' + kubectl_bin apply -f - +++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.XTKoPqpQt1 ++ local LAST_OUT=/tmp/tmp.CQeTBBKXxt + yq w - spec.parameters.pg-cluster cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + yq w - spec.namespace version-service-20560 +++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.x2DQZpW9Zy + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ local LAST_ERR=/tmp/tmp.R7DjewJsxJ ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.CQeTBBKXxt ++ cat /tmp/tmp.R7DjewJsxJ ++ rm /tmp/tmp.CQeTBBKXxt /tmp/tmp.R7DjewJsxJ ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-5f647d77b5-g5xnt + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.XTKoPqpQt1 pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.x2DQZpW9Zy + rm /tmp/tmp.XTKoPqpQt1 /tmp/tmp.x2DQZpW9Zy + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 latest 12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=12 + sort -s -u + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' ++ mktemp + local LAST_OUT=/tmp/tmp.815y8LzxGD ++ mktemp + local LAST_ERR=/tmp/tmp.wwAszyCkEo + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.815y8LzxGD + cat /tmp/tmp.wwAszyCkEo + rm /tmp/tmp.815y8LzxGD /tmp/tmp.wwAszyCkEo + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-latest-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster with major version latest release channel' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster with major version latest release channel ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + kubectl_bin -n version-service-20560 apply -f - + yq w - metadata.name cluster1-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.BH29LwMsK4 ++ mktemp + local LAST_ERR=/tmp/tmp.odGQSBpQkc + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + yq w - stringData.some-name some-name_pass + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.BH29LwMsK4 secret/cluster1-users-secret created + cat /tmp/tmp.odGQSBpQkc + rm /tmp/tmp.BH29LwMsK4 /tmp/tmp.odGQSBpQkc + return 0 + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.Lg6N13CLlJ ++ mktemp + local LAST_ERR=/tmp/tmp.ScOZ0x8dBc + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Lg6N13CLlJ secret/cluster1-pmm-secret configured + cat /tmp/tmp.ScOZ0x8dBc + rm /tmp/tmp.Lg6N13CLlJ /tmp/tmp.ScOZ0x8dBc + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.Ad75gbOlxB ++ mktemp + local LAST_ERR=/tmp/tmp.qY7pE5uzLB + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Ad75gbOlxB deployment.apps/pg-client unchanged + cat /tmp/tmp.qY7pE5uzLB + rm /tmp/tmp.Ad75gbOlxB /tmp/tmp.qY7pE5uzLB + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.name cluster1 + yq w - spec.clustername cluster1 + yq w - spec.database cluster1 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + yq w - spec.secretsName cluster1-users-secret + yq w - spec.keepData false + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + yq w - spec.user some-name + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + yq w - spec.name cluster1 + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.keepBackups false + yq w - metadata.labels.deployment-name cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.EocI33vxQN ++ mktemp + local LAST_ERR=/tmp/tmp.L5MHdlIoRl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.EocI33vxQN perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.L5MHdlIoRl + rm /tmp/tmp.EocI33vxQN /tmp/tmp.L5MHdlIoRl + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1....... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KLnRAZU6Hf +++ mktemp ++ local LAST_ERR=/tmp/tmp.WWZxGyo2lN ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.KLnRAZU6Hf ++ cat /tmp/tmp.WWZxGyo2lN ++ rm /tmp/tmp.KLnRAZU6Hf /tmp/tmp.WWZxGyo2lN ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.7CeMEhe2A3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yzinkZLhq0 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.7CeMEhe2A3 +++ cat /tmp/tmp.yzinkZLhq0 +++ rm /tmp/tmp.7CeMEhe2A3 /tmp/tmp.yzinkZLhq0 +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dIQn1W9ZSN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.eLEcAtldEG +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.dIQn1W9ZSN +++ cat /tmp/tmp.eLEcAtldEG +++ rm /tmp/tmp.dIQn1W9ZSN /tmp/tmp.eLEcAtldEG +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.NxjlD9LbpE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.K550lTa1Uz +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.NxjlD9LbpE +++ cat /tmp/tmp.K550lTa1Uz +++ rm /tmp/tmp.NxjlD9LbpE /tmp/tmp.K550lTa1Uz +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.JdhJa2fpiP ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0TsGz7hgxt +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.JdhJa2fpiP +++ cat /tmp/tmp.0TsGz7hgxt +++ rm /tmp/tmp.JdhJa2fpiP /tmp/tmp.0TsGz7hgxt +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.jdKhg04ygm +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.JHaFoy0ZWX ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.jdKhg04ygm ++++ cat /tmp/tmp.JHaFoy0ZWX ++++ rm /tmp/tmp.jdKhg04ygm /tmp/tmp.JHaFoy0ZWX ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ZdbbNKJqH2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.psjBL2cSLl +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ZdbbNKJqH2 +++ cat /tmp/tmp.psjBL2cSLl +++ rm /tmp/tmp.ZdbbNKJqH2 /tmp/tmp.psjBL2cSLl +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ywG9Gsy20b +++ mktemp ++ local LAST_ERR=/tmp/tmp.nVMXafAJ9a ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.ywG9Gsy20b ++ cat /tmp/tmp.nVMXafAJ9a ++ rm /tmp/tmp.ywG9Gsy20b /tmp/tmp.nVMXafAJ9a ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.ypTuEa2hBK ++ mktemp + local LAST_ERR=/tmp/tmp.5gHAZehUKD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ypTuEa2hBK + cat /tmp/tmp.5gHAZehUKD + rm /tmp/tmp.ypTuEa2hBK /tmp/tmp.5gHAZehUKD + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.00HGmLtnf5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NdCvq9I8fa +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.00HGmLtnf5 +++ cat /tmp/tmp.NdCvq9I8fa +++ rm /tmp/tmp.00HGmLtnf5 /tmp/tmp.NdCvq9I8fa +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qoYjtUa1X0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Vorc3qGSc6 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.qoYjtUa1X0 +++ cat /tmp/tmp.Vorc3qGSc6 +++ rm /tmp/tmp.qoYjtUa1X0 /tmp/tmp.Vorc3qGSc6 +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.cBXvXRvz5M ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1jPzKv2fIS +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.cBXvXRvz5M +++ cat /tmp/tmp.1jPzKv2fIS +++ rm /tmp/tmp.cBXvXRvz5M /tmp/tmp.1jPzKv2fIS +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.7VK6t7HqCp +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.HnRec5mit2 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.7VK6t7HqCp ++++ cat /tmp/tmp.HnRec5mit2 ++++ rm /tmp/tmp.7VK6t7HqCp /tmp/tmp.HnRec5mit2 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.d99bMp1pIh ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0NMrbFVDQ0 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.d99bMp1pIh +++ cat /tmp/tmp.0NMrbFVDQ0 +++ rm /tmp/tmp.d99bMp1pIh /tmp/tmp.0NMrbFVDQ0 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.22A2SzKyor +++ mktemp ++ local LAST_ERR=/tmp/tmp.3A6WaKgDib ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.22A2SzKyor ++ cat /tmp/tmp.3A6WaKgDib ++ rm /tmp/tmp.22A2SzKyor /tmp/tmp.3A6WaKgDib ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.51As5nJDUh ++ mktemp + local LAST_ERR=/tmp/tmp.SmmQ4OhZ0N + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.51As5nJDUh + cat /tmp/tmp.SmmQ4OhZ0N + rm /tmp/tmp.51As5nJDUh /tmp/tmp.SmmQ4OhZ0N + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.EvO1VdKTyq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.oUuEqOyiVt +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.EvO1VdKTyq +++ cat /tmp/tmp.oUuEqOyiVt +++ rm /tmp/tmp.EvO1VdKTyq /tmp/tmp.oUuEqOyiVt +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Ral8tQ4ht3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zzZ59yQe6X +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Ral8tQ4ht3 +++ cat /tmp/tmp.zzZ59yQe6X +++ rm /tmp/tmp.Ral8tQ4ht3 /tmp/tmp.zzZ59yQe6X +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.2m4O2vHj3X ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9u9VDCcSfa +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.2m4O2vHj3X +++ cat /tmp/tmp.9u9VDCcSfa +++ rm /tmp/tmp.2m4O2vHj3X /tmp/tmp.9u9VDCcSfa +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.95SMsbNvEX +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.o1Ht0nI7Eu ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.95SMsbNvEX ++++ cat /tmp/tmp.o1Ht0nI7Eu ++++ rm /tmp/tmp.95SMsbNvEX /tmp/tmp.o1Ht0nI7Eu ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Wtgq8cVAki ++++ mktemp +++ local LAST_ERR=/tmp/tmp.HoL8d5LxdM +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Wtgq8cVAki +++ cat /tmp/tmp.HoL8d5LxdM +++ rm /tmp/tmp.Wtgq8cVAki /tmp/tmp.HoL8d5LxdM +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GWIpl1bZRB +++ mktemp ++ local LAST_ERR=/tmp/tmp.sNAtFPykWM ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.GWIpl1bZRB ++ cat /tmp/tmp.sNAtFPykWM ++ rm /tmp/tmp.GWIpl1bZRB /tmp/tmp.sNAtFPykWM ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.rEm75P62TA ++ mktemp + local LAST_ERR=/tmp/tmp.77pdP9joeR + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.rEm75P62TA + cat /tmp/tmp.77pdP9joeR + rm /tmp/tmp.rEm75P62TA /tmp/tmp.77pdP9joeR + return 0 + sleep 10 + apply_version cluster1 9.9.9 12-latest 12 http://version-service:11000 3 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-latest + local db_version=12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=3 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.RKvVyymRJ7 ++ mktemp + local LAST_ERR=/tmp/tmp.tc6Rpzf5cs + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"12-latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.RKvVyymRJ7 perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.tc6Rpzf5cs + rm /tmp/tmp.RKvVyymRJ7 /tmp/tmp.tc6Rpzf5cs + return 0 + sleep 10 + wait_smart_update cluster1 3 + local cluster=cluster1 + local target=3 + set +o xtrace ........................................................ + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.UGRYPBDgRl ++ mktemp + local LAST_ERR=/tmp/tmp.MGGI7uq1yh + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.UGRYPBDgRl job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.MGGI7uq1yh + rm /tmp/tmp.UGRYPBDgRl /tmp/tmp.MGGI7uq1yh + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - spec.name fulldata-cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + yq w - spec.namespace version-service-20560 + yq w - spec.parameters.pg-cluster cluster1 ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tTKtQuP5ao +++ mktemp + kubectl_bin apply -f - + yq w - spec.parameters.job-name fulldata-cluster1 ++ local LAST_ERR=/tmp/tmp.IH8X71D0XA ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit + yq w - metadata.name fulldata-cluster1 +++ seq 0 2 ++ mktemp + local LAST_OUT=/tmp/tmp.4c6eHq7Z7K ++ mktemp + local LAST_ERR=/tmp/tmp.ezaAVE1Qqh + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.tTKtQuP5ao ++ cat /tmp/tmp.IH8X71D0XA ++ rm /tmp/tmp.tTKtQuP5ao /tmp/tmp.IH8X71D0XA ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-5f647d77b5-t4ljf + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.4c6eHq7Z7K pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.ezaAVE1Qqh + rm /tmp/tmp.4c6eHq7Z7K /tmp/tmp.ezaAVE1Qqh + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 12-latest 12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=12-latest + local db_version=12 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.0VR4YMwKKS ++ mktemp + local LAST_ERR=/tmp/tmp.e2rGlzrLbR + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.0VR4YMwKKS + cat /tmp/tmp.e2rGlzrLbR + rm /tmp/tmp.0VR4YMwKKS /tmp/tmp.e2rGlzrLbR + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-12-latest-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster update with explicitly specified image inside version service' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster update with explicitly specified image inside version service ----------------------------------------------------------------------------------- + PG_VER=13 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.jeVZkVLb5i ++ mktemp + local LAST_ERR=/tmp/tmp.vLx9NvZBry + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.jeVZkVLb5i deployment.apps/pg-client unchanged + cat /tmp/tmp.vLx9NvZBry + rm /tmp/tmp.jeVZkVLb5i /tmp/tmp.vLx9NvZBry + return 0 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-20560 apply -f - + yq w - metadata.name cluster1-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.gPxZb4SgTj ++ mktemp + local LAST_ERR=/tmp/tmp.Qw311hqYVy + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.gPxZb4SgTj secret/cluster1-users-secret created + cat /tmp/tmp.Qw311hqYVy + rm /tmp/tmp.gPxZb4SgTj /tmp/tmp.Qw311hqYVy + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.pTBDUDSLrb ++ mktemp + local LAST_ERR=/tmp/tmp.v0AY6IshFy + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.pTBDUDSLrb secret/cluster1-pmm-secret configured + cat /tmp/tmp.v0AY6IshFy + rm /tmp/tmp.pTBDUDSLrb /tmp/tmp.v0AY6IshFy + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.CeJHh1CPgC ++ mktemp + local LAST_ERR=/tmp/tmp.iEVTuR5I1g + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.CeJHh1CPgC deployment.apps/pg-client unchanged + cat /tmp/tmp.iEVTuR5I1g + rm /tmp/tmp.CeJHh1CPgC /tmp/tmp.iEVTuR5I1g + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - spec.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.keepData false + yq w - spec.clustername cluster1 + yq w - spec.user some-name + yq w - metadata.name cluster1 + yq w - spec.database cluster1 + yq w - spec.secretsName cluster1-users-secret + yq w - spec.keepBackups false + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + yq w - metadata.labels.name cluster1 + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.xV1eJsP7lH ++ mktemp + local LAST_ERR=/tmp/tmp.jA7Xf7i20X + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xV1eJsP7lH perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.jA7Xf7i20X + rm /tmp/tmp.xV1eJsP7lH /tmp/tmp.jA7Xf7i20X + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1......... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mDsGqlB4RX +++ mktemp ++ local LAST_ERR=/tmp/tmp.6YkQCxzrR0 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.mDsGqlB4RX ++ cat /tmp/tmp.6YkQCxzrR0 ++ rm /tmp/tmp.mDsGqlB4RX /tmp/tmp.6YkQCxzrR0 ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.8CGR0OV4IJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Seh8eyIIZc +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.8CGR0OV4IJ +++ cat /tmp/tmp.Seh8eyIIZc +++ rm /tmp/tmp.8CGR0OV4IJ /tmp/tmp.Seh8eyIIZc +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Hy4iMBSGBl ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Ck0GkegLvJ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Hy4iMBSGBl +++ cat /tmp/tmp.Ck0GkegLvJ +++ rm /tmp/tmp.Hy4iMBSGBl /tmp/tmp.Ck0GkegLvJ +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wnGJhLsIDT ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZBxaHJSUtR +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.wnGJhLsIDT +++ cat /tmp/tmp.ZBxaHJSUtR +++ rm /tmp/tmp.wnGJhLsIDT /tmp/tmp.ZBxaHJSUtR +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SBSg29OTzo ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VHcmFjkGzf +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.SBSg29OTzo +++ cat /tmp/tmp.VHcmFjkGzf +++ rm /tmp/tmp.SBSg29OTzo /tmp/tmp.VHcmFjkGzf +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.gAVXUlkmUx +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Nhcd4hKGKy ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.gAVXUlkmUx ++++ cat /tmp/tmp.Nhcd4hKGKy ++++ rm /tmp/tmp.gAVXUlkmUx /tmp/tmp.Nhcd4hKGKy ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.JyA1xlC3ES ++++ mktemp +++ local LAST_ERR=/tmp/tmp.X1HqgWh5Ur +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.JyA1xlC3ES +++ cat /tmp/tmp.X1HqgWh5Ur +++ rm /tmp/tmp.JyA1xlC3ES /tmp/tmp.X1HqgWh5Ur +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qeLVLJ9Dmw +++ mktemp ++ local LAST_ERR=/tmp/tmp.n94NYze6fR ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.qeLVLJ9Dmw ++ cat /tmp/tmp.n94NYze6fR ++ rm /tmp/tmp.qeLVLJ9Dmw /tmp/tmp.n94NYze6fR ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.yqucP6Gudl ++ mktemp + local LAST_ERR=/tmp/tmp.5tL3kgQsfW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.yqucP6Gudl + cat /tmp/tmp.5tL3kgQsfW + rm /tmp/tmp.yqucP6Gudl /tmp/tmp.5tL3kgQsfW + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.4hAydUCsVR ++++ mktemp +++ local LAST_ERR=/tmp/tmp.h6dnyyxrNr +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.4hAydUCsVR +++ cat /tmp/tmp.h6dnyyxrNr +++ rm /tmp/tmp.4hAydUCsVR /tmp/tmp.h6dnyyxrNr +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Uw90Q0w4DI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ojCG2Y32sj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Uw90Q0w4DI +++ cat /tmp/tmp.ojCG2Y32sj +++ rm /tmp/tmp.Uw90Q0w4DI /tmp/tmp.ojCG2Y32sj +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MIsqCWUbgm ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ROveHqWS3m +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MIsqCWUbgm +++ cat /tmp/tmp.ROveHqWS3m +++ rm /tmp/tmp.MIsqCWUbgm /tmp/tmp.ROveHqWS3m +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.5KBHntCQrj +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.HB8MxgSCgd ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.5KBHntCQrj ++++ cat /tmp/tmp.HB8MxgSCgd ++++ rm /tmp/tmp.5KBHntCQrj /tmp/tmp.HB8MxgSCgd ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.f9LgT7JUiq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1MgQdp6vXS +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.f9LgT7JUiq +++ cat /tmp/tmp.1MgQdp6vXS +++ rm /tmp/tmp.f9LgT7JUiq /tmp/tmp.1MgQdp6vXS +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2vyFiPv58E +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q2DZP2C5jX ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.2vyFiPv58E ++ cat /tmp/tmp.Q2DZP2C5jX ++ rm /tmp/tmp.2vyFiPv58E /tmp/tmp.Q2DZP2C5jX ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.8Y5kEFY1p6 ++ mktemp + local LAST_ERR=/tmp/tmp.cuN1QEn23r + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.8Y5kEFY1p6 + cat /tmp/tmp.cuN1QEn23r + rm /tmp/tmp.8Y5kEFY1p6 /tmp/tmp.cuN1QEn23r + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.b6gnH6ui2Y ++++ mktemp +++ local LAST_ERR=/tmp/tmp.QRBlqQ9BLy +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.b6gnH6ui2Y +++ cat /tmp/tmp.QRBlqQ9BLy +++ rm /tmp/tmp.b6gnH6ui2Y /tmp/tmp.QRBlqQ9BLy +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SW4otge3bM ++++ mktemp +++ local LAST_ERR=/tmp/tmp.DT98re3TLn +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.SW4otge3bM +++ cat /tmp/tmp.DT98re3TLn +++ rm /tmp/tmp.SW4otge3bM /tmp/tmp.DT98re3TLn +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.d0qIX57yVd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.4z6MDZkvqm +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.d0qIX57yVd +++ cat /tmp/tmp.4z6MDZkvqm +++ rm /tmp/tmp.d0qIX57yVd /tmp/tmp.4z6MDZkvqm +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Kg3tailSno +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.6wZSGh90z9 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.Kg3tailSno ++++ cat /tmp/tmp.6wZSGh90z9 ++++ rm /tmp/tmp.Kg3tailSno /tmp/tmp.6wZSGh90z9 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.fg2jNvmsJy ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gstGKbViJZ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.fg2jNvmsJy +++ cat /tmp/tmp.gstGKbViJZ +++ rm /tmp/tmp.fg2jNvmsJy /tmp/tmp.gstGKbViJZ +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fqrwR1mDJV +++ mktemp ++ local LAST_ERR=/tmp/tmp.m9OVld0k0b ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.fqrwR1mDJV ++ cat /tmp/tmp.m9OVld0k0b ++ rm /tmp/tmp.fqrwR1mDJV /tmp/tmp.m9OVld0k0b ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.9Ii7NN9M0R ++ mktemp + local LAST_ERR=/tmp/tmp.XUfubXDKVX + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9Ii7NN9M0R + cat /tmp/tmp.XUfubXDKVX + rm /tmp/tmp.9Ii7NN9M0R /tmp/tmp.XUfubXDKVX + return 0 + sleep 10 + apply_version cluster1 9.9.9 13.12 13 http://version-service:11000 4 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=13.12 + local db_version=13 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=4 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"13.12"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.Ad92ZCeqSL ++ mktemp + local LAST_ERR=/tmp/tmp.i0fq40C4cD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"13.12"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Ad92ZCeqSL perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.i0fq40C4cD + rm /tmp/tmp.Ad92ZCeqSL /tmp/tmp.i0fq40C4cD + return 0 + sleep 10 + wait_smart_update cluster1 4 + local cluster=cluster1 + local target=4 + set +o xtrace ..................................................... + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1 + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.COsceNQwun ++ mktemp + local LAST_ERR=/tmp/tmp.zn2vU2ZcnT + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.COsceNQwun job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.zn2vU2ZcnT + rm /tmp/tmp.COsceNQwun /tmp/tmp.zn2vU2ZcnT + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.namespace version-service-20560 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w - spec.parameters.pg-cluster cluster1 + yq w --style=single -- - spec.parameters.backrest-opts --type=full + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.9EnmmFMGCG ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ mktemp +++ mktemp + local LAST_ERR=/tmp/tmp.ZRYw04D6lH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ local LAST_OUT=/tmp/tmp.E9vvJ6vtet + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - +++ mktemp ++ local LAST_ERR=/tmp/tmp.8ptoLfuijV ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.E9vvJ6vtet ++ cat /tmp/tmp.8ptoLfuijV ++ rm /tmp/tmp.E9vvJ6vtet /tmp/tmp.8ptoLfuijV ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-57f668d478-x4rln + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9EnmmFMGCG pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.ZRYw04D6lH + rm /tmp/tmp.9EnmmFMGCG /tmp/tmp.ZRYw04D6lH + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 13.12 13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=13.12 + local db_version=13 + sort -s -u + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' ++ mktemp + local LAST_OUT=/tmp/tmp.XdSWTuvIui ++ mktemp + local LAST_ERR=/tmp/tmp.ggxotlupdH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.XdSWTuvIui + cat /tmp/tmp.ggxotlupdH + rm /tmp/tmp.XdSWTuvIui /tmp/tmp.ggxotlupdH + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-13.12-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'PG cluster with version service offline' + set +o xtrace ----------------------------------------------------------------------------------- PG cluster with version service offline ----------------------------------------------------------------------------------- + PG_VER=13 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name cluster1-users-secret + kubectl_bin -n version-service-20560 apply -f - + yq w - stringData.some-name some-name_pass ++ mktemp + local LAST_OUT=/tmp/tmp.JdIgiyfaVA ++ mktemp + local LAST_ERR=/tmp/tmp.pOJ2SAvor7 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.JdIgiyfaVA secret/cluster1-users-secret created + cat /tmp/tmp.pOJ2SAvor7 + rm /tmp/tmp.JdIgiyfaVA /tmp/tmp.pOJ2SAvor7 + return 0 + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.J8KzClv4FF ++ mktemp + local LAST_ERR=/tmp/tmp.moxSINzIV6 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.J8KzClv4FF secret/cluster1-pmm-secret configured + cat /tmp/tmp.moxSINzIV6 + rm /tmp/tmp.J8KzClv4FF /tmp/tmp.moxSINzIV6 + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.vFiO83yE7l ++ mktemp + local LAST_ERR=/tmp/tmp.3Ppdg4vhcB + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.vFiO83yE7l deployment.apps/pg-client unchanged + cat /tmp/tmp.3Ppdg4vhcB + rm /tmp/tmp.vFiO83yE7l /tmp/tmp.3Ppdg4vhcB + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - spec.clustername cluster1 + yq w - spec.name cluster1 + yq w - spec.database cluster1 + yq w - spec.keepBackups false + yq w - spec.user some-name + yq w - spec.keepData false + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger + yq w - metadata.labels.name cluster1 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo + yq w - spec.secretsName cluster1-users-secret + yq w - metadata.name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.lPN0ujjoWd ++ mktemp + local LAST_ERR=/tmp/tmp.4KnD7q4jpZ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.lPN0ujjoWd perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.4KnD7q4jpZ + rm /tmp/tmp.lPN0ujjoWd /tmp/tmp.4KnD7q4jpZ + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1........ + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7UsP4yWgNk +++ mktemp ++ local LAST_ERR=/tmp/tmp.kxS6kAQoIm ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.7UsP4yWgNk ++ cat /tmp/tmp.kxS6kAQoIm ++ rm /tmp/tmp.7UsP4yWgNk /tmp/tmp.kxS6kAQoIm ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dt24I5VMW5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CjRJOdl8pK +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.dt24I5VMW5 +++ cat /tmp/tmp.CjRJOdl8pK +++ rm /tmp/tmp.dt24I5VMW5 /tmp/tmp.CjRJOdl8pK +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.zHtrI0fHN8 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.WrNTPzusVx +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.zHtrI0fHN8 +++ cat /tmp/tmp.WrNTPzusVx +++ rm /tmp/tmp.zHtrI0fHN8 /tmp/tmp.WrNTPzusVx +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.kBLYjKZlqf ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NNJmPun6MU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.kBLYjKZlqf +++ cat /tmp/tmp.NNJmPun6MU +++ rm /tmp/tmp.kBLYjKZlqf /tmp/tmp.NNJmPun6MU +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tO3121iXVb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jWDq38fof4 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tO3121iXVb +++ cat /tmp/tmp.jWDq38fof4 +++ rm /tmp/tmp.tO3121iXVb /tmp/tmp.jWDq38fof4 +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.4DP2j592wL +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.mTGcmUhbCc ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.4DP2j592wL ++++ cat /tmp/tmp.mTGcmUhbCc ++++ rm /tmp/tmp.4DP2j592wL /tmp/tmp.mTGcmUhbCc ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tLvy666EYv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NcBPYnrpCz +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tLvy666EYv +++ cat /tmp/tmp.NcBPYnrpCz +++ rm /tmp/tmp.tLvy666EYv /tmp/tmp.NcBPYnrpCz +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ttJbVJtmcV +++ mktemp ++ local LAST_ERR=/tmp/tmp.9YzbWpuj37 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.ttJbVJtmcV ++ cat /tmp/tmp.9YzbWpuj37 ++ rm /tmp/tmp.ttJbVJtmcV /tmp/tmp.9YzbWpuj37 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.LcJJaT0xZS ++ mktemp + local LAST_ERR=/tmp/tmp.4KvYGbtjOB + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.LcJJaT0xZS + cat /tmp/tmp.4KvYGbtjOB + rm /tmp/tmp.LcJJaT0xZS /tmp/tmp.4KvYGbtjOB + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.9WzQogNbUH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.iWNKDg7GHk +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.9WzQogNbUH +++ cat /tmp/tmp.iWNKDg7GHk +++ rm /tmp/tmp.9WzQogNbUH /tmp/tmp.iWNKDg7GHk +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.YA3X3tggGD ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XbPv5hiZKH +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.YA3X3tggGD +++ cat /tmp/tmp.XbPv5hiZKH +++ rm /tmp/tmp.YA3X3tggGD /tmp/tmp.XbPv5hiZKH +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.QlSd9Y8dTc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.x7P3VjJXTY +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.QlSd9Y8dTc +++ cat /tmp/tmp.x7P3VjJXTY +++ rm /tmp/tmp.QlSd9Y8dTc /tmp/tmp.x7P3VjJXTY +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Nz5Zh0pZSE +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.SJU4KGB43o ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.Nz5Zh0pZSE ++++ cat /tmp/tmp.SJU4KGB43o ++++ rm /tmp/tmp.Nz5Zh0pZSE /tmp/tmp.SJU4KGB43o ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wMxhscogI4 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TmR48EaOFU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.wMxhscogI4 +++ cat /tmp/tmp.TmR48EaOFU +++ rm /tmp/tmp.wMxhscogI4 /tmp/tmp.TmR48EaOFU +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DN1Nu1bLTE +++ mktemp ++ local LAST_ERR=/tmp/tmp.WudFyfUD7n ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.DN1Nu1bLTE ++ cat /tmp/tmp.WudFyfUD7n ++ rm /tmp/tmp.DN1Nu1bLTE /tmp/tmp.WudFyfUD7n ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.ySLqGJoIKp ++ mktemp + local LAST_ERR=/tmp/tmp.RpY8vKxgdG + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ySLqGJoIKp + cat /tmp/tmp.RpY8vKxgdG + rm /tmp/tmp.ySLqGJoIKp /tmp/tmp.RpY8vKxgdG + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.75gkoYmuYw ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VdMmcZ557p +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.75gkoYmuYw +++ cat /tmp/tmp.VdMmcZ557p +++ rm /tmp/tmp.75gkoYmuYw /tmp/tmp.VdMmcZ557p +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.TKVq7tw7Q6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.obpwoSByHK +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.TKVq7tw7Q6 +++ cat /tmp/tmp.obpwoSByHK +++ rm /tmp/tmp.TKVq7tw7Q6 /tmp/tmp.obpwoSByHK +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.jFmoBDQ1MZ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Qqven9SSZd +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.jFmoBDQ1MZ +++ cat /tmp/tmp.Qqven9SSZd +++ rm /tmp/tmp.jFmoBDQ1MZ /tmp/tmp.Qqven9SSZd +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.2jOSZjKoAW +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.n1VavkRMJp ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.2jOSZjKoAW ++++ cat /tmp/tmp.n1VavkRMJp ++++ rm /tmp/tmp.2jOSZjKoAW /tmp/tmp.n1VavkRMJp ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OEuQiYaUhN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.u5GgcBzoqx +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.OEuQiYaUhN +++ cat /tmp/tmp.u5GgcBzoqx +++ rm /tmp/tmp.OEuQiYaUhN /tmp/tmp.u5GgcBzoqx +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZHUDQ5kczU +++ mktemp ++ local LAST_ERR=/tmp/tmp.OUwlvTMpbJ ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.ZHUDQ5kczU ++ cat /tmp/tmp.OUwlvTMpbJ ++ rm /tmp/tmp.ZHUDQ5kczU /tmp/tmp.OUwlvTMpbJ ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.JOjap0o3Fl ++ mktemp + local LAST_ERR=/tmp/tmp.8bLPc7yR7e + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.JOjap0o3Fl + cat /tmp/tmp.8bLPc7yR7e + rm /tmp/tmp.JOjap0o3Fl /tmp/tmp.8bLPc7yR7e + return 0 + sleep 10 + apply_version cluster1 9.9.9 latest 1.5.1-13 http://not-a-version-service/ 4 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=1.5.1-13 + local vs_endpoint=http://not-a-version-service/ + local update_finished_log_entries=4 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://not-a-version-service/","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.eP9H8780sP ++ mktemp + local LAST_ERR=/tmp/tmp.oJQ6ncVv20 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://not-a-version-service/","apply":"latest"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.eP9H8780sP perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.oJQ6ncVv20 + rm /tmp/tmp.eP9H8780sP /tmp/tmp.oJQ6ncVv20 + return 0 + sleep 10 + wait_smart_update cluster1 4 + local cluster=cluster1 + local target=4 + set +o xtrace + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1........ + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.9eEuXDNke4 ++ mktemp + local LAST_ERR=/tmp/tmp.rHIJLMDm0x + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9eEuXDNke4 job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.rHIJLMDm0x + rm /tmp/tmp.9eEuXDNke4 /tmp/tmp.rHIJLMDm0x + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - metadata.name fulldata-cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w - spec.namespace version-service-20560 + kubectl_bin apply -f - + yq w --style=single -- - spec.parameters.backrest-opts --type=full ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ mktemp +++ mktemp + yq w - spec.parameters.pg-cluster cluster1 + local LAST_OUT=/tmp/tmp.OtP3PQ4b8I ++ mktemp + local LAST_ERR=/tmp/tmp.kSBbVA1DFZ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ local LAST_OUT=/tmp/tmp.nM1LDYVlou + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - +++ mktemp ++ local LAST_ERR=/tmp/tmp.7gNVUhyiyu ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.nM1LDYVlou ++ cat /tmp/tmp.7gNVUhyiyu ++ rm /tmp/tmp.nM1LDYVlou /tmp/tmp.7gNVUhyiyu ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-65857c588c-fxb56 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.OtP3PQ4b8I pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.kSBbVA1DFZ + rm /tmp/tmp.OtP3PQ4b8I /tmp/tmp.kSBbVA1DFZ + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 latest 1.5.1-13 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=latest + local db_version=1.5.1-13 + sort -s -u + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' ++ mktemp + local LAST_OUT=/tmp/tmp.8b5I52Ls0Q ++ mktemp + local LAST_ERR=/tmp/tmp.fLMGIeInjv + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.8b5I52Ls0Q + cat /tmp/tmp.fLMGIeInjv + rm /tmp/tmp.8b5I52Ls0Q /tmp/tmp.fLMGIeInjv + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-latest-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/1.5.1-13.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Updating PG cluster with version service available but disabled' + set +o xtrace ----------------------------------------------------------------------------------- Updating PG cluster with version service available but disabled ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-20560 apply -f - + yq w - metadata.name cluster1-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.nPQraRCje9 ++ mktemp + local LAST_ERR=/tmp/tmp.LDZtTzkLh5 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nPQraRCje9 secret/cluster1-users-secret created + cat /tmp/tmp.LDZtTzkLh5 + rm /tmp/tmp.nPQraRCje9 /tmp/tmp.LDZtTzkLh5 + return 0 + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.tMQi9ekD5R ++ mktemp + local LAST_ERR=/tmp/tmp.Gnx27phP3w + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.tMQi9ekD5R secret/cluster1-pmm-secret configured + cat /tmp/tmp.Gnx27phP3w + rm /tmp/tmp.tMQi9ekD5R /tmp/tmp.Gnx27phP3w + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.22DGlKvne3 ++ mktemp + local LAST_ERR=/tmp/tmp.U6e00JKHx6 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.22DGlKvne3 deployment.apps/pg-client unchanged + cat /tmp/tmp.U6e00JKHx6 + rm /tmp/tmp.22DGlKvne3 /tmp/tmp.U6e00JKHx6 + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.deployment-name cluster1 + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.name cluster1 + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + yq w - spec.clustername cluster1 + yq w - spec.keepData false + yq w - spec.keepBackups false + yq w - metadata.labels.name cluster1 + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + yq w - spec.database cluster1 + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + yq w - spec.name cluster1 + yq w - spec.secretsName cluster1-users-secret + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.user some-name + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.nHIm8WgVVl ++ mktemp + local LAST_ERR=/tmp/tmp.zxfqzQjs37 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.nHIm8WgVVl perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.zxfqzQjs37 + rm /tmp/tmp.nHIm8WgVVl /tmp/tmp.zxfqzQjs37 + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1........ + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' cluster1-stanza-create+ retry=0 ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KRxbYBfHEP +++ mktemp ++ local LAST_ERR=/tmp/tmp.5JH3Ih0Iq8 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.KRxbYBfHEP ++ cat /tmp/tmp.5JH3Ih0Iq8 ++ rm /tmp/tmp.KRxbYBfHEP /tmp/tmp.5JH3Ih0Iq8 ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.f2CwIF5EBj ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Iwun2bf0E4 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.f2CwIF5EBj +++ cat /tmp/tmp.Iwun2bf0E4 +++ rm /tmp/tmp.f2CwIF5EBj /tmp/tmp.Iwun2bf0E4 +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ITn9ZVlZkH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.e0qWHc8Gro +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ITn9ZVlZkH +++ cat /tmp/tmp.e0qWHc8Gro +++ rm /tmp/tmp.ITn9ZVlZkH /tmp/tmp.e0qWHc8Gro +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tZ6zdCtL1B ++++ mktemp +++ local LAST_ERR=/tmp/tmp.rjkgYPyqwP +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tZ6zdCtL1B +++ cat /tmp/tmp.rjkgYPyqwP +++ rm /tmp/tmp.tZ6zdCtL1B /tmp/tmp.rjkgYPyqwP +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.KR1avcDA3r ++++ mktemp +++ local LAST_ERR=/tmp/tmp.YppHYEHzKR +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.KR1avcDA3r +++ cat /tmp/tmp.YppHYEHzKR +++ rm /tmp/tmp.KR1avcDA3r /tmp/tmp.YppHYEHzKR +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.HOCRha65Wi ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VfhMwcqYvx +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.HOCRha65Wi +++ cat /tmp/tmp.VfhMwcqYvx +++ rm /tmp/tmp.HOCRha65Wi /tmp/tmp.VfhMwcqYvx +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.yVmJG0tTdB +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.r4uZ7OyNHC ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.yVmJG0tTdB ++++ cat /tmp/tmp.r4uZ7OyNHC ++++ rm /tmp/tmp.yVmJG0tTdB /tmp/tmp.r4uZ7OyNHC ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ybkM3j6hRX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.P15kfqlQrU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ybkM3j6hRX +++ cat /tmp/tmp.P15kfqlQrU +++ rm /tmp/tmp.ybkM3j6hRX /tmp/tmp.P15kfqlQrU +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fzrX8jvJFT +++ mktemp ++ local LAST_ERR=/tmp/tmp.aVvlKCsIp6 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.fzrX8jvJFT ++ cat /tmp/tmp.aVvlKCsIp6 ++ rm /tmp/tmp.fzrX8jvJFT /tmp/tmp.aVvlKCsIp6 ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.TQPT1rRzji ++ mktemp + local LAST_ERR=/tmp/tmp.P58RQB2eLM + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.TQPT1rRzji + cat /tmp/tmp.P58RQB2eLM + rm /tmp/tmp.TQPT1rRzji /tmp/tmp.P58RQB2eLM + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ox1uyMyjcc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.nONvlvzqj7 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ox1uyMyjcc +++ cat /tmp/tmp.nONvlvzqj7 +++ rm /tmp/tmp.ox1uyMyjcc /tmp/tmp.nONvlvzqj7 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0chL5f995L ++++ mktemp +++ local LAST_ERR=/tmp/tmp.rIl5IgSaMe +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.0chL5f995L +++ cat /tmp/tmp.rIl5IgSaMe +++ rm /tmp/tmp.0chL5f995L /tmp/tmp.rIl5IgSaMe +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rpfeeb9PBn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uVzQiLXr59 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rpfeeb9PBn +++ cat /tmp/tmp.uVzQiLXr59 +++ rm /tmp/tmp.rpfeeb9PBn /tmp/tmp.uVzQiLXr59 +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.BzR5Idg1P3 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.DLYZJC7nhD ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.BzR5Idg1P3 ++++ cat /tmp/tmp.DLYZJC7nhD ++++ rm /tmp/tmp.BzR5Idg1P3 /tmp/tmp.DLYZJC7nhD ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Yi1lWzInNc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.emV6wCpXTK +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Yi1lWzInNc +++ cat /tmp/tmp.emV6wCpXTK +++ rm /tmp/tmp.Yi1lWzInNc /tmp/tmp.emV6wCpXTK +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RpX9Q5Pb5C +++ mktemp ++ local LAST_ERR=/tmp/tmp.jDmmFFuycp ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.RpX9Q5Pb5C ++ cat /tmp/tmp.jDmmFFuycp ++ rm /tmp/tmp.RpX9Q5Pb5C /tmp/tmp.jDmmFFuycp ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.3c3bf9NFAU ++ mktemp + local LAST_ERR=/tmp/tmp.KHUCtnIDSg + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.3c3bf9NFAU + cat /tmp/tmp.KHUCtnIDSg + rm /tmp/tmp.3c3bf9NFAU /tmp/tmp.KHUCtnIDSg + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MwCoQhoPvG ++++ mktemp +++ local LAST_ERR=/tmp/tmp.b0sERP7D38 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MwCoQhoPvG +++ cat /tmp/tmp.b0sERP7D38 +++ rm /tmp/tmp.MwCoQhoPvG /tmp/tmp.b0sERP7D38 +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.7k1yRYkf4O ++++ mktemp +++ local LAST_ERR=/tmp/tmp.CY7maa7P1e +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.7k1yRYkf4O +++ cat /tmp/tmp.CY7maa7P1e +++ rm /tmp/tmp.7k1yRYkf4O /tmp/tmp.CY7maa7P1e +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dcUAHHdLv9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.2Cg7iV9okS +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.dcUAHHdLv9 +++ cat /tmp/tmp.2Cg7iV9okS +++ rm /tmp/tmp.dcUAHHdLv9 /tmp/tmp.2Cg7iV9okS +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.8Oplzwu5vF +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.70csnPJhCC ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.8Oplzwu5vF ++++ cat /tmp/tmp.70csnPJhCC ++++ rm /tmp/tmp.8Oplzwu5vF /tmp/tmp.70csnPJhCC ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rG6HkjlzF3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wJzjmvNq4q +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.rG6HkjlzF3 +++ cat /tmp/tmp.wJzjmvNq4q +++ rm /tmp/tmp.rG6HkjlzF3 /tmp/tmp.wJzjmvNq4q +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PkVKh7oKl8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VXCkPmEenK ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.PkVKh7oKl8 ++ cat /tmp/tmp.VXCkPmEenK ++ rm /tmp/tmp.PkVKh7oKl8 /tmp/tmp.VXCkPmEenK ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.YdE3lSSj38 ++ mktemp + local LAST_ERR=/tmp/tmp.ACMQ3aJuOT + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.YdE3lSSj38 + cat /tmp/tmp.ACMQ3aJuOT + rm /tmp/tmp.YdE3lSSj38 /tmp/tmp.ACMQ3aJuOT + return 0 + sleep 10 + apply_version cluster1 9.9.9 disabled 1.5.1-12 http://version-service:11000 4 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=disabled + local db_version=1.5.1-12 + local vs_endpoint=http://version-service:11000 + local update_finished_log_entries=4 + kubectl_bin patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"disabled"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.F5YgPBsec1 ++ mktemp + local LAST_ERR=/tmp/tmp.4ZvK7y4hiB + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl patch perconapgcluster/cluster1 --type json '-p=[ {"op":"add","path":"/spec/upgradeOptions","value":{"versionServiceEndpoint":"http://version-service:11000","apply":"disabled"}}, {"op":"replace","path":"/metadata/labels/pgo-version","value":"9.9.9"}, {"op":"replace","path":"/spec/userLabels/pgo-version","value":"9.9.9"} ]' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.F5YgPBsec1 perconapgcluster.pg.percona.com/cluster1 patched + cat /tmp/tmp.4ZvK7y4hiB + rm /tmp/tmp.F5YgPBsec1 /tmp/tmp.4ZvK7y4hiB + return 0 + sleep 10 + wait_smart_update cluster1 4 + local cluster=cluster1 + local target=4 + set +o xtrace + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1........... + kubectl_bin delete jobs --selector=pg-cluster=cluster1 ++ mktemp + local LAST_OUT=/tmp/tmp.0aFlN4jxuc ++ mktemp + local LAST_ERR=/tmp/tmp.RT1fiM3LLr + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete jobs --selector=pg-cluster=cluster1 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.0aFlN4jxuc job.batch "backrest-backup-cluster1" deleted + cat /tmp/tmp.RT1fiM3LLr + rm /tmp/tmp.0aFlN4jxuc /tmp/tmp.RT1fiM3LLr + return 0 + create_backup cluster1 fulldata full + local cluster=cluster1 + local bckp_prefix=fulldata + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/backup.yml metadata.labels.pg-cluster cluster1 + yq w - spec.name fulldata-cluster1 + yq w - spec.namespace version-service-20560 + yq w - metadata.name fulldata-cluster1 + yq w - spec.parameters.job-name fulldata-cluster1 + yq w - spec.parameters.pg-cluster cluster1 + kubectl_bin apply -f - ++ mktemp + yq w --style=single -- - spec.parameters.backrest-opts --type=full ++ kubectl_bin get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' + local LAST_OUT=/tmp/tmp.B58DV5deMc +++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.DVRuICk9ki + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f - ++ local LAST_OUT=/tmp/tmp.pgKv2DSrRP +++ mktemp ++ local LAST_ERR=/tmp/tmp.5WtZaXx0Gm ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=cluster1-backrest-shared-repo,pg-cluster=cluster1 -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.pgKv2DSrRP ++ cat /tmp/tmp.5WtZaXx0Gm ++ rm /tmp/tmp.pgKv2DSrRP /tmp/tmp.5WtZaXx0Gm ++ return 0 + yq w - spec.parameters.podname cluster1-backrest-shared-repo-85c794b5cb-wbshm + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.B58DV5deMc pgtask.pg.percona.com/fulldata-cluster1 created + cat /tmp/tmp.DVRuICk9ki + rm /tmp/tmp.B58DV5deMc /tmp/tmp.DVRuICk9ki + return 0 + sleep 10 + wait_job_completion fulldata-cluster1 + local job=fulldata-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' fulldata-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/fulldata-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + verify_applied_version cluster1 9.9.9 disabled 1.5.1-12 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=disabled + local db_version=1.5.1-12 + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + sort -s -u ++ mktemp + local LAST_OUT=/tmp/tmp.WnsSTJVyBR ++ mktemp + local LAST_ERR=/tmp/tmp.AqgTDdt90D + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.WnsSTJVyBR + cat /tmp/tmp.AqgTDdt90D + rm /tmp/tmp.WnsSTJVyBR /tmp/tmp.AqgTDdt90D + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-disabled-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/1.5.1-12.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ...Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + desc 'Start the cluster up with version service images' + set +o xtrace ----------------------------------------------------------------------------------- Start the cluster up with version service images ----------------------------------------------------------------------------------- + PG_VER=12 + IMAGE_PGBOUNCER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + IMAGE_PG_HA=percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + IMAGE_BACKREST=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + IMAGE_BACKREST_REPO=percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + IMAGE_PGBADGER=percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + yq w -i /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.labels.pgo-version 9.9.9 + yq w -i /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml spec.userLabels.pgo-version 9.9.9 + yq w -i /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml spec.upgradeOptions.versionServiceEndpoint http://version-service:11000 + yq w -i --style=single /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml spec.upgradeOptions.apply recommended + spinup_pgcluster cluster1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local cluster=cluster1 + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + local namespace=version-service-20560 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets cluster1 version-service-20560 + local cluster=cluster1 + local namespace=version-service-20560 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + kubectl_bin -n version-service-20560 apply -f - + yq w - metadata.name cluster1-users-secret ++ mktemp + local LAST_OUT=/tmp/tmp.IkTL2Gj7Zf ++ mktemp + local LAST_ERR=/tmp/tmp.qQ53KN229i + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.IkTL2Gj7Zf secret/cluster1-users-secret created + cat /tmp/tmp.qQ53KN229i + rm /tmp/tmp.IkTL2Gj7Zf /tmp/tmp.qQ53KN229i + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-723/deploy/users-secret.yaml + yq w - metadata.name cluster1-pmm-secret + kubectl_bin -n version-service-20560 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.Xsxu72m8pe ++ mktemp + local LAST_ERR=/tmp/tmp.UGOS9cRAsk + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Xsxu72m8pe secret/cluster1-pmm-secret configured + cat /tmp/tmp.UGOS9cRAsk + rm /tmp/tmp.Xsxu72m8pe /tmp/tmp.UGOS9cRAsk + return 0 + kubectl_bin -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.Kooz7l4APF ++ mktemp + local LAST_ERR=/tmp/tmp.xAxI3z8Jqa + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Kooz7l4APF deployment.apps/pg-client unchanged + cat /tmp/tmp.xAxI3z8Jqa + rm /tmp/tmp.Kooz7l4APF /tmp/tmp.xAxI3z8Jqa + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml cluster1 false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml + local name=cluster1 + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/conf/1.5.1_cr.yaml metadata.annotations.current-primary cluster1 + yq w - metadata.labels.crunchy-pgha-scope cluster1 + yq w - metadata.labels.name cluster1 + yq w - metadata.labels.pg-cluster cluster1 + yq w - metadata.name cluster1 + yq w - spec.database cluster1 + yq w - spec.name cluster1 + yq w - spec.keepData false + yq w - spec.pmm.image perconalab/pmm-client:2.41.0 + yq w - spec.backup.backrestRepoImage percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo + yq w - spec.pgBadger.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger + yq w - spec.pgPrimary.image percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha + yq w - spec.backup.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest + yq w - spec.pgBouncer.image percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer + yq w - spec.clustername cluster1 + yq w - spec.keepBackups false + yq w - spec.secretsName cluster1-users-secret + yq w - spec.user some-name + yq w - metadata.labels.deployment-name cluster1 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 ++ mktemp + local LAST_OUT=/tmp/tmp.3BKRtCILDE ++ mktemp + local LAST_ERR=/tmp/tmp.dUAJpcsO9f + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/cr.yaml -n version-service-20560 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.3BKRtCILDE perconapgcluster.pg.percona.com/cluster1 created + cat /tmp/tmp.dUAJpcsO9f + rm /tmp/tmp.3BKRtCILDE /tmp/tmp.dUAJpcsO9f + return 0 + wait_deployment cluster1-backrest-shared-repo + local name=cluster1-backrest-shared-repo + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1-backrest-shared-repo.. + wait_deployment cluster1 + local name=cluster1 + local target_namespace=version-service-20560 + sleep 10 + set +o xtrace cluster1........... + [[ false == \g\c\s\+* ]] + wait_job_completion cluster1-stanza-create false + local job=cluster1-stanza-create + local mandatory=false + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 2 -ge 120 ]] + [[ 2 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 3 -ge 120 ]] + [[ 3 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 4 -ge 120 ]] + [[ 4 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 5 -ge 120 ]] + [[ 5 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 6 -ge 120 ]] + [[ 6 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 7 -ge 120 ]] + [[ 7 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 8 -ge 120 ]] + [[ 8 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 9 -ge 120 ]] + [[ 9 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 10 -ge 120 ]] + [[ 10 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 11 -ge 120 ]] + [[ 11 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 12 -ge 120 ]] + [[ 12 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 13 -ge 120 ]] + [[ 13 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 14 -ge 120 ]] + [[ 14 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 15 -ge 120 ]] + [[ 15 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 16 -ge 120 ]] + [[ 16 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 17 -ge 120 ]] + [[ 17 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 18 -ge 120 ]] + [[ 18 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 19 -ge 120 ]] + [[ 19 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 20 -ge 120 ]] + [[ 20 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 21 -ge 120 ]] + [[ 21 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 22 -ge 120 ]] + [[ 22 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 23 -ge 120 ]] + [[ 23 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 24 -ge 120 ]] + [[ 24 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 25 -ge 120 ]] + [[ 25 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 26 -ge 120 ]] + [[ 26 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 27 -ge 120 ]] + [[ 27 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 28 -ge 120 ]] + [[ 28 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 29 -ge 120 ]] + [[ 29 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 30 -ge 120 ]] + [[ 30 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 31 -ge 120 ]] + [[ 31 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 32 -ge 120 ]] + [[ 32 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 33 -ge 120 ]] + [[ 33 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 34 -ge 120 ]] + [[ 34 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 35 -ge 120 ]] + [[ 35 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 36 -ge 120 ]] + [[ 36 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 37 -ge 120 ]] + [[ 37 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 38 -ge 120 ]] + [[ 38 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 39 -ge 120 ]] + [[ 39 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 40 -ge 120 ]] + [[ 40 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 41 -ge 120 ]] + [[ 41 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 42 -ge 120 ]] + [[ 42 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 43 -ge 120 ]] + [[ 43 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 44 -ge 120 ]] + [[ 44 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 45 -ge 120 ]] + [[ 45 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 46 -ge 120 ]] + [[ 46 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 47 -ge 120 ]] + [[ 47 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 48 -ge 120 ]] + [[ 48 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 49 -ge 120 ]] + [[ 49 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 50 -ge 120 ]] + [[ 50 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 51 -ge 120 ]] + [[ 51 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 52 -ge 120 ]] + [[ 52 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 53 -ge 120 ]] + [[ 53 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 54 -ge 120 ]] + [[ 54 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 55 -ge 120 ]] + [[ 55 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 56 -ge 120 ]] + [[ 56 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 57 -ge 120 ]] + [[ 57 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 58 -ge 120 ]] + [[ 58 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 59 -ge 120 ]] + [[ 59 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 60 -ge 120 ]] + [[ 60 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 61 -ge 120 ]] + [[ 61 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 62 -ge 120 ]] + [[ 62 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 63 -ge 120 ]] + [[ 63 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 64 -ge 120 ]] + [[ 64 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 65 -ge 120 ]] + [[ 65 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 66 -ge 120 ]] + [[ 66 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 67 -ge 120 ]] + [[ 67 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 68 -ge 120 ]] + [[ 68 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 69 -ge 120 ]] + [[ 69 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 70 -ge 120 ]] + [[ 70 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 71 -ge 120 ]] + [[ 71 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 72 -ge 120 ]] + [[ 72 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 73 -ge 120 ]] + [[ 73 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 74 -ge 120 ]] + [[ 74 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 75 -ge 120 ]] + [[ 75 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 76 -ge 120 ]] + [[ 76 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 77 -ge 120 ]] + [[ 77 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 78 -ge 120 ]] + [[ 78 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 79 -ge 120 ]] + [[ 79 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 80 -ge 120 ]] + [[ 80 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 81 -ge 120 ]] + [[ 81 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 82 -ge 120 ]] + [[ 82 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 83 -ge 120 ]] + [[ 83 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 84 -ge 120 ]] + [[ 84 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 85 -ge 120 ]] + [[ 85 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 86 -ge 120 ]] + [[ 86 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 87 -ge 120 ]] + [[ 87 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 88 -ge 120 ]] + [[ 88 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 89 -ge 120 ]] + [[ 89 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 90 -ge 120 ]] + [[ 90 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 91 -ge 120 ]] + [[ 91 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 92 -ge 120 ]] + [[ 92 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 93 -ge 120 ]] + [[ 93 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 94 -ge 120 ]] + [[ 94 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 95 -ge 120 ]] + [[ 95 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 96 -ge 120 ]] + [[ 96 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 97 -ge 120 ]] + [[ 97 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 98 -ge 120 ]] + [[ 98 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 99 -ge 120 ]] + [[ 99 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 100 -ge 120 ]] + [[ 100 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 101 -ge 120 ]] + [[ 101 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 102 -ge 120 ]] + [[ 102 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 103 -ge 120 ]] + [[ 103 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 104 -ge 120 ]] + [[ 104 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 105 -ge 120 ]] + [[ 105 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 106 -ge 120 ]] + [[ 106 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 107 -ge 120 ]] + [[ 107 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 108 -ge 120 ]] + [[ 108 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 109 -ge 120 ]] + [[ 109 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 110 -ge 120 ]] + [[ 110 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 111 -ge 120 ]] + [[ 111 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 112 -ge 120 ]] + [[ 112 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 113 -ge 120 ]] + [[ 113 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 114 -ge 120 ]] + [[ 114 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 115 -ge 120 ]] + [[ 115 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 116 -ge 120 ]] + [[ 116 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 117 -ge 120 ]] + [[ 117 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 118 -ge 120 ]] + [[ 118 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 119 -ge 120 ]] + [[ 119 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/cluster1-stanza-create -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 120 -ge 120 ]] + [[ false == \t\r\u\e ]] + [[ 120 -ge 120 ]] + [[ false != \t\r\u\e ]] + echo Can not detect job. Passing by. Can not detect job. Passing by. + set -o xtrace + return 0 + wait_job_completion backrest-backup-cluster1 + local job=backrest-backup-cluster1 + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' backrest-backup-cluster1+ retry=0 ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/backrest-backup-cluster1 -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency cluster1 + cluster_name=cluster1 + wait_cluster_status cluster1 'pgcluster Initialized' + local cluster_name=cluster1 + local 'target_status=pgcluster Initialized' + local target_namespace=version-service-20560 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LuBt0RV9nl +++ mktemp ++ local LAST_ERR=/tmp/tmp.VgNK6wUaK7 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.LuBt0RV9nl ++ cat /tmp/tmp.VgNK6wUaK7 ++ rm /tmp/tmp.LuBt0RV9nl /tmp/tmp.VgNK6wUaK7 ++ return 0 + status_query='kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hGRhZYhqCG ++++ mktemp +++ local LAST_ERR=/tmp/tmp.PAyb19AwQ8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.hGRhZYhqCG +++ cat /tmp/tmp.PAyb19AwQ8 +++ rm /tmp/tmp.hGRhZYhqCG /tmp/tmp.PAyb19AwQ8 +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.fKbhxmN2KL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wfSaNd1pbm +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.fKbhxmN2KL +++ cat /tmp/tmp.wfSaNd1pbm +++ rm /tmp/tmp.fKbhxmN2KL /tmp/tmp.wfSaNd1pbm +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.G1wxO1IkIc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ep8qYrqn7a +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.G1wxO1IkIc +++ cat /tmp/tmp.ep8qYrqn7a +++ rm /tmp/tmp.G1wxO1IkIc /tmp/tmp.ep8qYrqn7a +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.HQYO9V7diq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bplPJci2YV +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.HQYO9V7diq +++ cat /tmp/tmp.bplPJci2YV +++ rm /tmp/tmp.HQYO9V7diq /tmp/tmp.bplPJci2YV +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.fVkjNWo5FW +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.wITmRrXZsV ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.fVkjNWo5FW ++++ cat /tmp/tmp.wITmRrXZsV ++++ rm /tmp/tmp.fVkjNWo5FW /tmp/tmp.wITmRrXZsV ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Po1TpDgfyM ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bJhNxaSCj3 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Po1TpDgfyM +++ cat /tmp/tmp.bJhNxaSCj3 +++ rm /tmp/tmp.Po1TpDgfyM /tmp/tmp.bJhNxaSCj3 +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@cluster1.version-service-20560 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z58383RuyE +++ mktemp ++ local LAST_ERR=/tmp/tmp.6HrKoTUZcq ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.Z58383RuyE ++ cat /tmp/tmp.6HrKoTUZcq ++ rm /tmp/tmp.Z58383RuyE /tmp/tmp.6HrKoTUZcq ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.GxqNeFb9Ec ++ mktemp + local LAST_ERR=/tmp/tmp.6EriJDKY6d + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.GxqNeFb9Ec + cat /tmp/tmp.6EriJDKY6d + rm /tmp/tmp.GxqNeFb9Ec /tmp/tmp.6EriJDKY6d + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.M33VaSo6jH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.4ri0wWI1bE +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.M33VaSo6jH +++ cat /tmp/tmp.4ri0wWI1bE +++ rm /tmp/tmp.M33VaSo6jH /tmp/tmp.4ri0wWI1bE +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.R5WWgg84jX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.FJQkjK9Zs2 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.R5WWgg84jX +++ cat /tmp/tmp.FJQkjK9Zs2 +++ rm /tmp/tmp.R5WWgg84jX /tmp/tmp.FJQkjK9Zs2 +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.X7o8AR3vdd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zJs8NYKiw6 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.X7o8AR3vdd +++ cat /tmp/tmp.zJs8NYKiw6 +++ rm /tmp/tmp.X7o8AR3vdd /tmp/tmp.zJs8NYKiw6 +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.w75UJ8HCkg +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.TbTbRb6Pnx ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.w75UJ8HCkg ++++ cat /tmp/tmp.TbTbRb6Pnx ++++ rm /tmp/tmp.w75UJ8HCkg /tmp/tmp.TbTbRb6Pnx ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.urBPU2daMB ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Qx7fb1qaGv +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.urBPU2daMB +++ cat /tmp/tmp.Qx7fb1qaGv +++ rm /tmp/tmp.urBPU2daMB /tmp/tmp.Qx7fb1qaGv +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8XholLehPs +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ctw2RbzLYF ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.8XholLehPs ++ cat /tmp/tmp.Ctw2RbzLYF ++ rm /tmp/tmp.8XholLehPs /tmp/tmp.Ctw2RbzLYF ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.eD26Jp7eJ9 ++ mktemp + local LAST_ERR=/tmp/tmp.YAlAilWhXA + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.eD26Jp7eJ9 + cat /tmp/tmp.YAlAilWhXA + rm /tmp/tmp.eD26Jp7eJ9 /tmp/tmp.YAlAilWhXA + return 0 ++ get_psql_user_pass postgres cluster1 ++ local user=postgres ++ local cluster=cluster1 ++ local namespace=version-service-20560 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DMoCc6cdi9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1iJkKNZfWp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.DMoCc6cdi9 +++ cat /tmp/tmp.1iJkKNZfWp +++ rm /tmp/tmp.DMoCc6cdi9 /tmp/tmp.1iJkKNZfWp +++ return 0 ++ [[ xcluster1 == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6buj5Pjog1 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.dxZNyU0rou +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl get perconapgcluster/cluster1 -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.6buj5Pjog1 +++ cat /tmp/tmp.dxZNyU0rou +++ rm /tmp/tmp.6buj5Pjog1 /tmp/tmp.dxZNyU0rou +++ return 0 +++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.UcwqH4b4s5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.PJSoob7JiA +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.UcwqH4b4s5 +++ cat /tmp/tmp.PJSoob7JiA +++ rm /tmp/tmp.UcwqH4b4s5 /tmp/tmp.PJSoob7JiA +++ return 0 ++ [[ xcluster1-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.meiezjwpL2 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.sgVqmhwNL0 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++++ kubectl -n version-service-20560 get perconapgcluster/cluster1 -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.meiezjwpL2 ++++ cat /tmp/tmp.sgVqmhwNL0 ++++ rm /tmp/tmp.meiezjwpL2 /tmp/tmp.sgVqmhwNL0 ++++ return 0 +++ kubectl_bin -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lvVGwxjic2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0ILBBaRbCz +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service +++ kubectl -n version-service-20560 get secret/cluster1-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.lvVGwxjic2 +++ cat /tmp/tmp.0ILBBaRbCz +++ rm /tmp/tmp.lvVGwxjic2 /tmp/tmp.0ILBBaRbCz +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@cluster1.version-service-20560 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@cluster1.version-service-20560 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lpnvupEKDl +++ mktemp ++ local LAST_ERR=/tmp/tmp.bDiUwIKv8U ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl -n version-service-20560 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.lpnvupEKDl ++ cat /tmp/tmp.bDiUwIKv8U ++ rm /tmp/tmp.lpnvupEKDl /tmp/tmp.bDiUwIKv8U ++ return 0 + local client_container=pg-client-76df9df876-9kdhm + kubectl_bin exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.wkGP3dGH5k ++ mktemp + local LAST_ERR=/tmp/tmp.UHqF7cIfR7 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl exec pg-client-76df9df876-9kdhm -n version-service-20560 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@cluster1.version-service-20560.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.wkGP3dGH5k + cat /tmp/tmp.UHqF7cIfR7 + rm /tmp/tmp.wkGP3dGH5k /tmp/tmp.UHqF7cIfR7 + return 0 + sleep 10 + verify_applied_version cluster1 9.9.9 recommended 14 + local cluster=cluster1 + local version_to_upgrade=9.9.9 + local release_channel=recommended + local db_version=14 + sort -s -u + kubectl_bin get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' ++ mktemp + local LAST_OUT=/tmp/tmp.SWikksCL0t ++ mktemp + local LAST_ERR=/tmp/tmp.QATsOWf3VG + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get deployments,jobs --selector=pg-cluster=cluster1 -o 'jsonpath={range .items[*]}{range .spec.template.spec.containers[*]}{.image}{"\n"}{end}{end}' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.SWikksCL0t + cat /tmp/tmp.QATsOWf3VG + rm /tmp/tmp.SWikksCL0t /tmp/tmp.QATsOWf3VG + return 0 + diff -u /tmp/tmp.09n2hxddFJ/9.9.9-deployed-recommended-images.txt /mnt/jenkins/workspace/cloud-pg-operator_PR-723/e2e-tests/version-service/compare/14.txt + kubectl delete perconapgcluster/cluster1 perconapgcluster.pg.percona.com "cluster1" deleted + wait_for_delete deployment/cluster1 + local res=deployment/cluster1 + set +o xtrace deployment/cluster1 - ....Error from server (NotFound): deployments.apps "cluster1" not found + wait_for_delete deployment/cluster1-backrest-shared-repo + local res=deployment/cluster1-backrest-shared-repo + set +o xtrace deployment/cluster1-backrest-shared-repo - Error from server (NotFound): deployments.apps "cluster1-backrest-shared-repo" not found + destroy version-service-20560 + grep -v level=info ++ get_operator_pod + grep -v level=debug ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' + grep -v 'Getting tasks for pod' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + sort -u + grep -v 'get backup status: Job.batch' + tee /tmp/tmp.09n2hxddFJ/operator.log +++ mktemp + grep -v 'the object has been modified' + grep -v 'Getting pods from source' ++ local LAST_OUT=/tmp/tmp.i8pACKmd5Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.8BjKJHeb3T ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.i8pACKmd5Q ++ cat /tmp/tmp.8BjKJHeb3T ++ rm /tmp/tmp.i8pACKmd5Q /tmp/tmp.8BjKJHeb3T ++ return 0 + kubectl_bin logs postgres-operator-7c96f98c7d-hrdfz -c operator ++ mktemp + local LAST_OUT=/tmp/tmp.C7tDlmcJbr ++ mktemp + local LAST_ERR=/tmp/tmp.vlQBW0f3vj + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl logs postgres-operator-7c96f98c7d-hrdfz -c operator + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.C7tDlmcJbr + cat /tmp/tmp.vlQBW0f3vj + rm /tmp/tmp.C7tDlmcJbr /tmp/tmp.vlQBW0f3vj + return 0 ] { } }, ] } }, { } }, ] { } }, { } ] } { } }, ] } ] { } }, }, { ] { } }, ] ] ], { } }, }, { }, { },{ }] }], ] { } }, , ,{ ], { } }, }], { } }, }], } ], { } }, } ] ], } } ], }, }, { } }, { } }, }, 2024/04/24 16:39:42 INF 1 (localhost:4150) connecting to nsqd 2024/04/24 16:40:44 INF 2 (localhost:4150) connecting to nsqd 2024/04/24 16:40:46 INF 3 (localhost:4150) connecting to nsqd 2024/04/24 16:41:05 INF 4 (localhost:4150) connecting to nsqd 2024/04/24 16:41:33 INF 5 (localhost:4150) connecting to nsqd 2024/04/24 16:41:33 INF 6 (localhost:4150) connecting to nsqd 2024/04/24 16:41:33 INF 7 (localhost:4150) connecting to nsqd 2024/04/24 16:50:34 INF 8 (localhost:4150) connecting to nsqd 2024/04/24 16:52:14 INF 9 (localhost:4150) connecting to nsqd 2024/04/24 16:52:46 INF 10 (localhost:4150) connecting to nsqd 2024/04/24 16:52:49 INF 11 (localhost:4150) connecting to nsqd 2024/04/24 16:53:12 INF 12 (localhost:4150) connecting to nsqd 2024/04/24 16:53:17 INF 13 (localhost:4150) connecting to nsqd 2024/04/24 16:53:33 INF 14 (localhost:4150) connecting to nsqd 2024/04/24 16:54:46 INF 15 (localhost:4150) connecting to nsqd 2024/04/24 16:54:49 INF 16 (localhost:4150) connecting to nsqd 2024/04/24 16:55:06 INF 17 (localhost:4150) connecting to nsqd 2024/04/24 16:55:35 INF 18 (localhost:4150) connecting to nsqd 2024/04/24 16:55:36 INF 19 (localhost:4150) connecting to nsqd 2024/04/24 16:55:36 INF 20 (localhost:4150) connecting to nsqd 2024/04/24 17:04:47 INF 21 (localhost:4150) connecting to nsqd 2024/04/24 17:06:23 INF 22 (localhost:4150) connecting to nsqd 2024/04/24 17:06:52 INF 23 (localhost:4150) connecting to nsqd 2024/04/24 17:06:55 INF 24 (localhost:4150) connecting to nsqd 2024/04/24 17:07:19 INF 25 (localhost:4150) connecting to nsqd 2024/04/24 17:07:24 INF 26 (localhost:4150) connecting to nsqd 2024/04/24 17:07:39 INF 27 (localhost:4150) connecting to nsqd 2024/04/24 17:08:21 INF 28 (localhost:4150) connecting to nsqd 2024/04/24 17:08:22 INF 29 (localhost:4150) connecting to nsqd 2024/04/24 17:08:29 INF 30 (localhost:4150) connecting to nsqd 2024/04/24 17:08:55 INF 31 (localhost:4150) connecting to nsqd 2024/04/24 17:08:56 INF 32 (localhost:4150) connecting to nsqd 2024/04/24 17:08:56 INF 33 (localhost:4150) connecting to nsqd 2024/04/24 17:16:02 INF 34 (localhost:4150) connecting to nsqd 2024/04/24 17:16:29 INF 35 (localhost:4150) connecting to nsqd 2024/04/24 17:16:32 INF 36 (localhost:4150) connecting to nsqd 2024/04/24 17:16:56 INF 37 (localhost:4150) connecting to nsqd 2024/04/24 17:17:01 INF 38 (localhost:4150) connecting to nsqd 2024/04/24 17:17:15 INF 39 (localhost:4150) connecting to nsqd 2024/04/24 17:18:01 INF 40 (localhost:4150) connecting to nsqd 2024/04/24 17:18:02 INF 41 (localhost:4150) connecting to nsqd 2024/04/24 17:18:09 INF 42 (localhost:4150) connecting to nsqd 2024/04/24 17:18:36 INF 43 (localhost:4150) connecting to nsqd 2024/04/24 17:18:36 INF 44 (localhost:4150) connecting to nsqd 2024/04/24 17:18:37 INF 45 (localhost:4150) connecting to nsqd 2024/04/24 17:22:39 INF 46 (localhost:4150) connecting to nsqd 2024/04/24 17:24:15 INF 47 (localhost:4150) connecting to nsqd 2024/04/24 17:24:43 INF 48 (localhost:4150) connecting to nsqd 2024/04/24 17:24:46 INF 49 (localhost:4150) connecting to nsqd 2024/04/24 17:25:10 INF 50 (localhost:4150) connecting to nsqd 2024/04/24 17:25:15 INF 51 (localhost:4150) connecting to nsqd 2024/04/24 17:25:31 INF 52 (localhost:4150) connecting to nsqd 2024/04/24 17:26:20 INF 53 (localhost:4150) connecting to nsqd 2024/04/24 17:26:21 INF 54 (localhost:4150) connecting to nsqd 2024/04/24 17:26:28 INF 55 (localhost:4150) connecting to nsqd 2024/04/24 17:26:56 INF 56 (localhost:4150) connecting to nsqd 2024/04/24 17:26:56 INF 57 (localhost:4150) connecting to nsqd 2024/04/24 17:26:56 INF 58 (localhost:4150) connecting to nsqd 2024/04/24 17:31:07 INF 59 (localhost:4150) connecting to nsqd 2024/04/24 17:32:37 INF 60 (localhost:4150) connecting to nsqd 2024/04/24 17:32:52 INF 61 (localhost:4150) connecting to nsqd 2024/04/24 17:32:54 INF 62 (localhost:4150) connecting to nsqd 2024/04/24 17:33:14 INF 63 (localhost:4150) connecting to nsqd 2024/04/24 17:33:19 INF 64 (localhost:4150) connecting to nsqd 2024/04/24 17:33:37 INF 65 (localhost:4150) connecting to nsqd 2024/04/24 17:34:28 INF 66 (localhost:4150) connecting to nsqd 2024/04/24 17:34:29 INF 67 (localhost:4150) connecting to nsqd 2024/04/24 17:34:37 INF 68 (localhost:4150) connecting to nsqd 2024/04/24 17:35:03 INF 69 (localhost:4150) connecting to nsqd 2024/04/24 17:35:03 INF 70 (localhost:4150) connecting to nsqd 2024/04/24 17:35:04 INF 71 (localhost:4150) connecting to nsqd 2024/04/24 17:42:32 INF 72 (localhost:4150) connecting to nsqd 2024/04/24 17:44:07 INF 73 (localhost:4150) connecting to nsqd 2024/04/24 17:44:31 INF 74 (localhost:4150) connecting to nsqd 2024/04/24 17:44:35 INF 75 (localhost:4150) connecting to nsqd 2024/04/24 17:44:54 INF 76 (localhost:4150) connecting to nsqd 2024/04/24 17:44:59 INF 77 (localhost:4150) connecting to nsqd 2024/04/24 17:45:16 INF 78 (localhost:4150) connecting to nsqd 2024/04/24 17:46:06 INF 79 (localhost:4150) connecting to nsqd 2024/04/24 17:46:07 INF 80 (localhost:4150) connecting to nsqd 2024/04/24 17:46:13 INF 81 (localhost:4150) connecting to nsqd 2024/04/24 17:46:40 INF 82 (localhost:4150) connecting to nsqd 2024/04/24 17:46:40 INF 83 (localhost:4150) connecting to nsqd 2024/04/24 17:46:40 INF 84 (localhost:4150) connecting to nsqd 2024/04/24 17:49:47 INF 85 (localhost:4150) connecting to nsqd 2024/04/24 17:50:02 INF 86 (localhost:4150) connecting to nsqd 2024/04/24 17:50:03 INF 87 (localhost:4150) connecting to nsqd 2024/04/24 17:50:22 INF 88 (localhost:4150) connecting to nsqd 2024/04/24 17:50:27 INF 89 (localhost:4150) connecting to nsqd 2024/04/24 17:50:42 INF 90 (localhost:4150) connecting to nsqd 2024/04/24 17:51:28 INF 91 (localhost:4150) connecting to nsqd 2024/04/24 17:51:29 INF 92 (localhost:4150) connecting to nsqd 2024/04/24 17:51:37 INF 93 (localhost:4150) connecting to nsqd 2024/04/24 17:52:04 INF 94 (localhost:4150) connecting to nsqd 2024/04/24 17:52:04 INF 95 (localhost:4150) connecting to nsqd 2024/04/24 17:52:05 INF 96 (localhost:4150) connecting to nsqd 2024/04/24 17:55:28 INF 97 (localhost:4150) connecting to nsqd 2024/04/24 17:55:53 INF 98 (localhost:4150) connecting to nsqd 2024/04/24 17:55:55 INF 99 (localhost:4150) connecting to nsqd 2024/04/24 17:56:13 INF 100 (localhost:4150) connecting to nsqd 2024/04/24 17:56:18 INF 101 (localhost:4150) connecting to nsqd 2024/04/24 17:56:37 INF 102 (localhost:4150) connecting to nsqd 2024/04/24 17:57:22 INF 103 (localhost:4150) connecting to nsqd 2024/04/24 17:57:23 INF 104 (localhost:4150) connecting to nsqd 2024/04/24 17:57:30 INF 105 (localhost:4150) connecting to nsqd 2024/04/24 17:57:57 INF 106 (localhost:4150) connecting to nsqd 2024/04/24 17:57:57 INF 107 (localhost:4150) connecting to nsqd 2024/04/24 17:57:57 INF 108 (localhost:4150) connecting to nsqd 2024/04/24 18:01:17 INF 109 (localhost:4150) connecting to nsqd 2024/04/24 18:01:40 INF 110 (localhost:4150) connecting to nsqd 2024/04/24 18:01:45 INF 111 (localhost:4150) connecting to nsqd "accessModes": [ "affinity": "affinity": { "affinity": { "podAntiAffinity": { "allowPrivilegeEscalation": false, "allowPrivilegeEscalation": false, "annotations": {"keep-backups":"false","keep-data":"false"}, "apiVersion": "apps/v1", "apiVersion": "batch/v1", "apiVersion": "v1", "apiVersion": "v1", "apiVersion": "v1", "backrest-command": "backup", "backrest-command": "stanza-create", "claimName": "cluster1-pgbr-repo" "cluster1" "command": [ "configMap": { "configMap": { "configMapKeyRef": { "containerName": "database", "containerPort": 10000, "containerPort": 2022, "containerPort": 5432, "containerPort": 8009, "containers": [ "containers": [{ "cpu": "1", "cpu": "200m", "cpu": "500m", "cpu": "500m", "crunchydata" "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin" "defaultMode": 288 "defaultMode": 288, "defaultMode": 288, "defaultMode": 420, "deployment-name": "cluster1" "deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1","name": "cluster1","pg-cluster": "cluster1" "deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1","name": "cluster1","pg-cluster": "cluster1" "deployment-name": "cluster1-repl1" "deployment-name": "cluster1-repl2" "deployment-name": "cluster1-repl2","pgo-version": "1.5.1","name": "cluster1-replica","pg-cluster": "cluster1" "deployment-name": "cluster1-repl2","pgo-version": "9.9.9","name": "cluster1-replica","pg-cluster": "cluster1" "divisor": "1m", "dnsPolicy": "ClusterFirst" "downwardAPI": { "emptyDir": { "emptyDir": { "emptyDir": { "emptyDir": { "env": [ "env": [{ "env": [ { "exec": { "fieldPath": "metadata.annotations" "fieldPath": "metadata.labels" "fieldPath": "metadata.labels['crunchy-pgha-scope']" "fieldPath": "metadata.namespace" "fieldRef": { "fieldRef": { "fsGroup": 2, I0424 16:39:21.606924 1 shared_informer.go:240] Waiting for caches to sync for namespace I0424 16:39:21.707311 1 shared_informer.go:247] Caches are synced for namespace I0424 16:39:21.831637 1 shared_informer.go:240] Waiting for caches to sync for version-service-20560 I0424 16:39:21.932084 1 shared_informer.go:247] Caches are synced for version-service-20560 "image": "perconalab/percona-postgresql-operator:main-ppg12-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg12-pgbadger", "image": "perconalab/percona-postgresql-operator:main-ppg13-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg13-pgbadger", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbadger", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbouncer", "image": "perconalab/percona-postgresql-operator:main-ppg14-postgres-ha", "image": "perconalab/percona-postgresql-operator:PR-723-7ed19e708-pgo-rmdata", "image": "percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest", "image": "percona/percona-postgresql-operator:1.5.1-ppg12-pgbackrest-repo", "image": "percona/percona-postgresql-operator:1.5.1-ppg12-pgbadger", "image": "percona/percona-postgresql-operator:1.5.1-ppg12-pgbouncer", "image": "percona/percona-postgresql-operator:1.5.1-ppg12-postgres-ha", "image": "percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest", "image": "percona/percona-postgresql-operator:1.5.1-ppg13-pgbackrest-repo", "image": "percona/percona-postgresql-operator:1.5.1-ppg13-pgbadger", "image": "percona/percona-postgresql-operator:1.5.1-ppg13-pgbouncer", "image": "percona/percona-postgresql-operator:1.5.1-ppg13-postgres-ha", "imagePullPolicy": "IfNotPresent" "imagePullPolicy": "IfNotPresent", "imagePullPolicy": "IfNotPresent", "initialDelaySeconds": 15 "initialDelaySeconds": 20, "initialDelaySeconds": 30, "items": [ "items": [ "items": [ "key": "config", "key": "config", "key": "crunchy-pgbouncer", "key": "init" "key": "password" "key": "pg-cluster", "key": "pgo-backrest", "key": "pgo-backrest-job", "key": "pgo-pg-database", "key": "pg-pod-anti-affinity", "key": "tls.crt", "key": "tls.key", "key": "vendor", "kind": "Deployment", "kind": "Job", "kind": "PersistentVolumeClaim", "kind": "Service", "labels": { "labels": { "labels": { "labelSelector": { "limits": { "livenessProbe": { "matchExpressions": [ "matchLabels": { "matchLabels": { "maxSurge": 1 "maxUnavailable": 1, "medium": "Memory", "medium": "Memory" "medium": "Memory", "medium": "Memory", "medium": "Memory", "memory": "128Mi" "memory": "256Mi" "memory": "48Mi" "memory": "64Mi" "metadata": { "metadata": { "metadata": { "mountPath": "/backrestrepo", "mountPath": "/dev/shm", "mountPath": "/etc/pgbackrest" "mountPath": "/etc/pgbackrest/conf.d", "mountPath": "/etc/podinfo", "mountPath": "/etc/ssh", "mountPath": "/etc/ssh", "mountPath": "/home/pgbackrest" "mountPath": "/pgconf", "mountPath": "/pgconf/", "mountPath": "/pgconf/pgreplicator", "mountPath": "/pgconf/pgsuper", "mountPath": "/pgconf/pguser", "mountPath": "/pgconf/tls", "mountPath": "/pgconf/tls/pgbouncer", "mountPath": "/pgconf/tls-replication", "mountPath": "/pgdata", "mountPath": "/pgdata", "mountPath": "/report", "mountPath": "/sshd", "mountPath": "/sshd", "mountPath": "/tmp", "mountPath": "/tmp", "mountPath": "/tmp" "mountPath": "/tmp", "name": "backrest", "name": "backrest-backup-cluster1", "name": "backrestrepo", "name": "backrestrepo", "name": "BACKREST_SKIP_CREATE_STANZA", "name": "BADGER_TARGET", "name": "cluster1", "name": "cluster1", "name": "cluster1-backrest-shared-repo", "name": "cluster1-backrest-shared-repo" "name": "cluster1-backrest-shared-repo", "name": "cluster1-backrest-shared-repo", "name": "cluster1-backrest-shared-repo", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer", "name": "cluster1-pgbouncer-cm" "name": "cluster1-pgbouncer-secret", "name": "cluster1-pgbouncer-secret", "name": "cluster1-pgbr-repo", "name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1" "name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1" "name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "9.9.9" "name": "cluster1-pgha-config", "name": "cluster1-pgha-config", "name": "cluster1-repl1", "name": "cluster1-repl1", "name": "cluster1-repl2", "name": "cluster1-repl2", "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1","pgo-version": "1.5.1" "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1","pgo-version": "1.5.1" "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.5.1" "name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.5.1" "name": "cluster1-replication-ssl-keypair", "name": "cluster1-rmdata-fpkv", "name": "cluster1-rmdata-fpkv", "name": "cluster1-rmdata-hfkq", "name": "cluster1-rmdata-hfkq", "name": "cluster1-rmdata-iiak", "name": "cluster1-rmdata-iiak", "name": "cluster1-rmdata-oecn", "name": "cluster1-rmdata-oecn", "name": "cluster1-rmdata-pbtk", "name": "cluster1-rmdata-pbtk", "name": "cluster1-rmdata-relz", "name": "cluster1-rmdata-relz", "name": "cluster1-rmdata-ujkr", "name": "cluster1-rmdata-ujkr", "name": "cluster1-rmdata-xers", "name": "cluster1-rmdata-xers", "name": "cluster1-rmdata-yrdk", "name": "cluster1-rmdata-yrdk", "name": "cluster1-ssl-ca" "name": "cluster1-ssl-ca" "name": "cluster1-ssl-keypair" "name": "cluster1-ssl-keypair" "name": "cluster1-stanza-create", "name": "COMMAND", "name": "COMMAND_OPTS", "name": "database", "name": "dshm" "name": "dshm", "name": "ENABLE_SSHD", "name": "fulldata-cluster1", "name": "IS_BACKUP", "name": "IS_REPLICA", "name": "LD_PRELOAD", "name": "LD_PRELOAD", "name": "MODE", "name": "NAMESPACE", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_PASSWD", "name": "NSS_WRAPPER_PASSWD", "name": "patroni" "name": "PATRONI_KUBERNETES_LABELS", "name": "PATRONI_KUBERNETES_NAMESPACE", "name": "PATRONI_KUBERNETES_SCOPE_LABEL", "name": "PATRONI_LOG_LEVEL", "name": "PATRONI_POSTGRESQL_DATA_DIR", "name": "PATRONI_SCOPE", "name": "pgbackrest" "name": "pgbackrest-conf", "name": "pgbackrest-conf", "name": "pgbackrest-config" "name": "pgbackrest-config", "name": "PGBACKREST_DB_HOST", "name": "PGBACKREST_DB_PATH", "name": "PGBACKREST_DB_PATH", "name": "pgbackrest-home", "name": "pgbackrest-home", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_REPO1_HOST", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_STANZA", "name": "PGBACKREST_STANZA", "name": "pgbadger" "name": "pgbadger", "name": "PGBADGER_SERVICE_PORT", "name": "pgbouncer" "name": "pgbouncer", "name": "pgbouncer-conf", "name": "pgbouncer-conf", "name": "PG_CLUSTER", "name": "pgconf-volume" "name": "pgconf-volume", "name": "pgdata", "name": "pgdata", "name": "pgdata", "name": "PGHA_DATABASE", "name": "PGHA_INIT", "name": "PGHA_PASSWORD_TYPE", "name": "PGHA_PGBACKREST", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_S3_VERIFY_TLS", "name": "PGHA_PG_PORT", "name": "PGHA_REPLICA_REINIT_ON_START_FAIL", "name": "PGHA_SCOPE", "name": "PGHA_STANDBY", "name": "PGHA_SYNC_REPLICATION", "name": "PGHA_TLS_ENABLED", "name": "PGHA_TLS_ONLY", "name": "PGHA_USER", "name": "PGHOST", "name": "PG_PASSWORD", "name": "PG_PRIMARY_SERVICE_NAME", "name": "PITR_TARGET", "name": "podinfo" "name": "podinfo", "name": "PODNAME", "name": "postgres" "name": "primary-volume" "name": "primary-volume", "name": "REMOVE_BACKUP", "name": "REMOVE_DATA", "name": "REPLICA_NAME", "name": "report", "name": "report", "name": "rmdata", "name": "root-volume" "name": "root-volume", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "sshd", "name": "sshd", "name": "sshd", "name": "sshd", "name": "tls-pgbouncer" "name": "tls-pgbouncer", "name": "tls-replication" "name": "tls-replication", "name": "tls-server" "name": "tls-server", "name": "tmp" "name": "tmp" "name": "tmp", "name": "tmp", "name": "tmp", "name": "tmp" "name": "user-volume" "name": "user-volume", "nodePort": 0 "operator": "Exists" "operator": "In", "/opt/crunchy/bin/postgres-ha/health/pgha-liveness.sh" "/opt/crunchy/bin/postgres-ha/health/pgha-readiness.sh" "optional": true "path": "annotations" "path": "cpu_limit", "path": "cpu_request", "path": "labels" "path": "mem_limit", "path": "mem_request", "path": "ssh_config" "path": "ssh_config" "path": "tls-replication.crt" "path": "tls-replication.key" "periodSeconds": 10 "periodSeconds": 15, "persistentVolumeClaim": { "persistentVolumeClaim":{"claimName":"cluster1"} "persistentVolumeClaim":{"claimName":"cluster1-repl1"} "persistentVolumeClaim":{"claimName":"cluster1-repl2"} "pg-cluster": "cluster1" "pg-cluster": "cluster1", "pg-cluster": "cluster1" "pg-cluster": "cluster1", "pg-cluster": "cluster1" "pg-cluster": "cluster1", "pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1","name": "cluster1" "pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1","name": "cluster1" "pg-cluster": "cluster1","deployment-name": "cluster1-repl1","pgo-version": "1.5.1","name": "cluster1-replica" "pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.5.1","name": "cluster1-replica" "pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "1.5.1","name": "cluster1-replica" "pg-cluster": "cluster1","deployment-name": "cluster1-repl2","pgo-version": "9.9.9","name": "cluster1-replica" "pgo-backrest-job": "true", "pgo-backrest-repo": "true" "pgo-backrest-repo": "true", "pgo-backrest-repo": "true" "pgo-backrest": "true", "pgo-pg-database": "true", "pgo-pg-database": "true", "pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1" "pgouser": "admin","crunchy-pgha-scope": "cluster1","pgo-version": "1.5.1","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1" "pgo-version": "1.5.1","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1" "pgo-version": "1.5.1","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1" "pgo-version": "1.5.1","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pgo-version": "1.5.1","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pgo-version": "1.5.1","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2" "pgo-version": "1.5.1","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl2" "pgo-version": "9.9.9","name": "cluster1","pg-cluster": "cluster1","deployment-name": "cluster1","pgouser": "admin","crunchy-pgha-scope": "cluster1" "pgo-version": "9.9.9","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pgo-version": "9.9.9","name": "cluster1-replica","pg-cluster": "cluster1","deployment-name": "cluster1-repl1" "pg-pod-anti-affinity": "preferred", "pg-pod-anti-affinity": "preferred", "pgrmdata": "true", "pgrmdata": "true", "podAffinityTerm": { "podAntiAffinity": { "podAntiAffinity": { "podAntiAffinity": { "port": 10000 "port": 2022, "ports": [{ "ports": [{ "ports": [ { "preferredDuringSchedulingIgnoredDuringExecution": [ "privileged": false, "privileged": false, "projected": { "projected": { "projected": { "projected": { "sources": [] } "protocol": "TCP", "protocol": "TCP", "protocol": "TCP", "readinessProbe": { "readinessProbe": { "readOnly": false "readOnly": false "readOnly": false "readOnly": false "readOnlyRootFilesystem": true "readOnlyRootFilesystem": true "readOnly": true "readOnly": true "readOnly": true "readOnly": true "ReadWriteOnce" "replicas": 0, "replicas": 1, "replicas": 3, "requests": { "requests": { "resourceFieldRef": { "resource": "limits.cpu" "resource": "limits.memory" "resource": "requests.cpu" "resource": "requests.memory" "resources": { "resources": { "resources": { "restartPolicy": "Always", "restartPolicy": "Never" "rollingUpdate": { "runAsNonRoot": true "secret": { "secret": { "secret": { "secret": { "secret": { "secretKeyRef": { "secretName": "cluster1-backrest-repo-config" "secretName": "cluster1-backrest-repo-config", "secretName": "cluster1-backrest-repo-config" "secretName": "cluster1-backrest-repo-config", "secretName": "cluster1-postgres-secret" "secretName": "cluster1-primaryuser-secret" "secretName": "cluster1-some-name-secret" "securityContext": { "securityContext": { "securityContext": {}, "securityContext": { "securityContext": {"fsGroup": 26,"supplementalGroups": [1001]}, "securityContext": {"runAsNonRoot":true,"fsGroup":26}, "selector": { "selector": { "serviceAccountName": "pgo-backrest", "serviceAccountName": "pgo-default", "serviceAccountName": "pgo-pg", "serviceAccountName": "pgo-target", "service-name": "cluster1-backrest-shared-repo", "service-name": "cluster1-backrest-shared-repo", "service-name": "cluster1-pgbouncer", "service-name": "cluster1-pgbouncer", "service-name": "cluster1-pgbouncer", "sessionAffinity": "None" "sizeLimit": "128Ki" "sizeLimit": "16Mi" "sizeLimit": "1Mi" "sizeLimit": "2Mi" "sizeLimit": "64Mi" "sizeLimit": "64Mi" "sources": [ "sources": [ "sources": [ "spec": { "spec": { "spec": { "storage": "1Gi" "strategy": { "targetPort": 2022, "tcpSocket": { "template": { time="2024-04-24T16:39:21Z" level=warning msg="PGAdminStorage setting not set, will use PrimaryStorage setting" func="github.com/percona/percona-postgresql-operator/internal/config.(*PgoConfig).Validate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/config/pgoconfig.go:337" version=1.6.0 time="2024-04-24T16:39:46Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T16:39:51Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T16:40:44Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T16:40:44Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T16:41:33Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T16:41:34Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T16:41:34Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T16:47:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T16:47:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T16:47:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T16:48:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T16:48:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T16:48:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T16:50:24Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.6.0 time="2024-04-24T16:50:34Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T16:50:34Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl1/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.6.0 time="2024-04-24T16:52:50Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:52:52Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:52:53Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:52:54Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:52:55Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:52:56Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:53:02Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:53:06Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:53:09Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:53:11Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:53:12Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T16:53:17Z" level=error msg="jobs.batch \"cluster1-rmdata-hfkq\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T16:53:38Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T16:53:43Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T16:54:46Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T16:54:46Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T16:55:36Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T16:55:36Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:01:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:01:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:01:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:02:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:02:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:02:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:04:33Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.6.0 time="2024-04-24T17:04:40Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:04:47Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:04:47Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl2/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.6.0 time="2024-04-24T17:06:55Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:06:57Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:06:58Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:06:59Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:00Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:07Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:11Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:12Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:16Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:18Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:19Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:20Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:07:24Z" level=error msg="jobs.batch \"cluster1-rmdata-fpkv\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T17:07:43Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:07:48Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:08:21Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T17:08:21Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T17:08:56Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:08:56Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:13:31Z" level=error msg="14-recommended value for spec.upgradeOptions.apply option is not supported" func="github.com/percona/percona-postgresql-operator/percona/controllers/version.EnsureVersion()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/version/version.go:48" version=1.6.0 time="2024-04-24T17:13:31Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.6.0 time="2024-04-24T17:14:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:14:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:14:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:15:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:15:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:15:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:16:32Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:33Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:36Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:37Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:38Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:39Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:44Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:48Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:49Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:53Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:55Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:16:56Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:17:01Z" level=error msg="jobs.batch \"cluster1-rmdata-yrdk\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T17:17:20Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:17:25Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:18:01Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T17:18:01Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T17:18:37Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:18:38Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:22:28Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.6.0 time="2024-04-24T17:22:39Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:22:39Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl2/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.6.0 time="2024-04-24T17:24:46Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:24:48Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:24:49Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:24:50Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:24:51Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:24:52Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:24:58Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:25:02Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:25:03Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:25:07Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:25:09Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:25:10Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:25:15Z" level=error msg="jobs.batch \"cluster1-rmdata-relz\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T17:25:35Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:25:40Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:26:20Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T17:26:20Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T17:26:56Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:26:56Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:28:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:28:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:28:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:30:46Z" level=error msg="command terminated with exit code 2" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:30:48Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.6.0 time="2024-04-24T17:30:59Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:31:06Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:31:06Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl1/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.6.0 time="2024-04-24T17:32:55Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:32:56Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:32:57Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:32:58Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:32:59Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:33:02Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:33:06Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:33:08Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:33:11Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:33:13Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:33:14Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:33:19Z" level=error msg="jobs.batch \"cluster1-rmdata-ujkr\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T17:33:42Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:33:47Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:34:28Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T17:34:28Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T17:35:04Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:35:04Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:42:18Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.6.0 time="2024-04-24T17:42:25Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:42:32Z" level=error msg="command terminated with exit code 1" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:42:32Z" level=warning msg="pg_ctl: PID file \"/pgdata/cluster1-repl1/postmaster.pid\" does not exist\nIs server running?\n" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.applyUpdateToPostgresInstance()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/rolling.go:177" version=1.6.0 time="2024-04-24T17:44:35Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:36Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:37Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:38Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:39Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:40Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:43Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:47Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:48Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:51Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:53Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:54Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:44:59Z" level=error msg="jobs.batch \"cluster1-rmdata-oecn\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T17:45:20Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:45:25Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:46:06Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T17:46:06Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T17:46:40Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:46:41Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:47:24Z" level=error msg="update perconapgcluster: ensure version failed to check version: version service apply: Get \"http://not-a-version-service/versions/v1/pg-operator/9.9.9/latest?customResourceUid=58e99074-d7d4-45d4-b95d-8b24c574b092&databaseVersion=13.12+-+Percona+Distribution\": dial tcp: lookup not-a-version-service on 10.65.128.10:53: no such host" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:501" version=1.6.0 time="2024-04-24T17:47:24Z" level=error msg="update perconapgcluster: scheduled update: add func to cron: empty spec string" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgc.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgc/pgc.go:509" version=1.6.0 time="2024-04-24T17:48:21Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:48:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:48:21Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:49:22Z" level=error msg="-{\"authentication\":{\"replication\":{\"sslcert\":\"/pgconf/tls-replication/tls.crt\",\"sslkey\":\"/pgconf/tls-replication/tls.key\",\"sslmode\":\"verify-ca\",\"sslrootcert\":\"/pgconf/tls/ca.crt\"}},\"callbacks\":{\"on_role_change\":\"/opt/crunchy/bin/postgres-ha/callbacks/pgha-on-role-change.sh\"},\"create_replica_methods\":[\"pgbackrest\",\"basebackup\"],\"pg_hba\":[\"local all postgres peer\",\"hostssl replication primaryuser 0.0.0.0/0 cert\",\"hostssl all primaryuser 0.0.0.0/0 reject\",\"hostssl all all 0.0.0.0/0 md5\",\"host replication primaryuser 0.0.0.0/0 md5\",\"host all primaryuser 0.0.0.0/0 reject\",\"host all all 0.0.0.0/0 md5\"],\"pgbackrest\":{\"command\":\"/opt/crunchy/bin/postgres-ha/pgbackrest/pgbackrest-create-replica.sh replica\",\"keep_data\":true,\"no_params\":true},\"pgbackrest_standby\":{\"command\":\"/opt/crunchy/bin/postgres-ha/pgbackrest/pgbackrest-create-replica.sh standby\",\"keep_data\":true,\"no_master\":1,\"no_params\":true},\"pgpass\":\"/tmp/.pgpass\",\"remove_data_directory_on_rewind_failure\":true,\"use_unix_socket\":true}\n+{\"authentication\":{\"replication\":{\"sslkey\":\"/pgconf/tls-replication/tls.key\"}},\"callbacks\":{\"on_role_change\":\"/opt/crunchy/bin/postgres-ha/callbacks/pgha-on-role-change.sh\"},\"create_replica_methods\":[\"pgbackrest\",\"basebackup\"],\"pgbackrest\":{\"command\":\"/opt/crunchy/bin/postgres-ha/pgbackrest/pgbackrest-create-replica.sh replica\",\"keep_data\":true,\"no_params\":true},\"pgbackrest_standby\":{\"command\":\"/opt/crunchy/bin/postgres-ha/pgbackrest/pgbackrest-create-replica.sh standby\",\"keep_data\":true,\"no_master\":1,\"no_params\":true},\"pgpass\":\"/tmp/.pgpass\",\"remove_data_directory_on_rewind_failure\":true,\"use_unix_socket\":true}\n\x1b[0;32mWed Apr 24 17:49:22 UTC 2024 INFO: Reload Config: Detected config change, reloading local configuration\x1b[0m\n\x1b[0;31mWed Apr 24 17:49:22 UTC 2024 ERROR: Error reloading local config: Unable to reload local configuration\x1b[0m\n" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:49:22Z" level=error msg="command terminated with exit code 7" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:49:22Z" level=error msg="command terminated with exit code 7" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:50:04Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:05Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:06Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:07Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:08Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:12Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:16Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:17Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:20Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:21Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:22Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:50:27Z" level=error msg="jobs.batch \"cluster1-rmdata-iiak\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T17:50:46Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:50:51Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:51:28Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T17:51:28Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T17:52:05Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:52:05Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:52:22Z" level=error msg="command terminated with exit code 7" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:52:22Z" level=error msg="command terminated with exit code 7" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:177" version=1.6.0 time="2024-04-24T17:52:22Z" level=error msg="-{}\n+{\"authentication\":{\"replication\":{\"sslcert\":\"/pgconf/tls-replication/tls.crt\",\"sslkey\":\"/pgconf/tls-replication/tls.key\",\"sslmode\":\"verify-ca\",\"sslrootcert\":\"/pgconf/tls/ca.crt\"}},\"callbacks\":{\"on_role_change\":\"/opt/crunchy/bin/postgres-ha/callbacks/pgha-on-role-change.sh\"},\"create_replica_methods\":[\"pgbackrest\",\"basebackup\"],\"pg_hba\":[\"local all postgres peer\",\"hostssl replication primaryuser 0.0.0.0/0 cert\",\"hostssl all primaryuser 0.0.0.0/0 reject\",\"hostssl all all 0.0.0.0/0 md5\",\"host replication primaryuser 0.0.0.0/0 md5\",\"host all primaryuser 0.0.0.0/0 reject\",\"host all all 0.0.0.0/0 md5\"],\"pgbackrest\":{\"command\":\"/opt/crunchy/bin/postgres-ha/pgbackrest/pgbackrest-create-replica.sh replica\",\"keep_data\":true,\"no_params\":true},\"pgbackrest_standby\":{\"command\":\"/opt/crunchy/bin/postgres-ha/pgbackrest/pgbackrest-create-replica.sh standby\",\"keep_data\":true,\"no_master\":1,\"no_params\":true},\"pgpass\":\"/tmp/.pgpass\",\"remove_data_directory_on_rewind_failure\":true,\"use_unix_socket\":true}\n\x1b[0;32mWed Apr 24 17:52:22 UTC 2024 INFO: Reload Config: Detected config change, reloading local configuration\x1b[0m\n\x1b[0;31mWed Apr 24 17:52:22 UTC 2024 ERROR: Error reloading local config: Unable to reload local configuration\x1b[0m\n" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).apply()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:255" version=1.6.0 time="2024-04-24T17:55:55Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:55:56Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:55:57Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:55:58Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:55:59Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:56:00Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:56:02Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:56:07Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:56:10Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:56:12Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:56:13Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T17:56:18Z" level=error msg="jobs.batch \"cluster1-rmdata-pbtk\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 time="2024-04-24T17:56:41Z" level=error msg="get deployment cluster1-repl1: deployments.apps \"cluster1-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:56:46Z" level=error msg="get deployment cluster1-repl2: deployments.apps \"cluster1-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.6.0 time="2024-04-24T17:57:22Z" level=error msg="Error completing workflow cluster1-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.6.0 time="2024-04-24T17:57:22Z" level=error msg="pgtasks.pg.percona.com \"cluster1-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.6.0 time="2024-04-24T17:57:57Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:57:58Z" level=warning msg="crv1 pgreplica cluster1-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:57:58Z" level=warning msg="crv1 pgreplica cluster1-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.6.0 time="2024-04-24T17:58:22Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.6.0 time="2024-04-24T17:58:22Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.6.0 time="2024-04-24T17:58:22Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.6.0 time="2024-04-24T18:01:17Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:18Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:19Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:20Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:21Z" level=error msg="ConfigMap Controller: cannot find pgcluster for configMap cluster1-pgha-config (namespace version-service-20560),ignoring" func="github.com/percona/percona-postgresql-operator/internal/controller/configmap.(*Controller).handleConfigMapSync()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/configmap/synchandler.go:55" version=1.6.0 time="2024-04-24T18:01:21Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:22Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:23Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:28Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:33Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:37Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:39Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:40Z" level=error msg="pgclusters.pg.percona.com \"cluster1\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.6.0 time="2024-04-24T18:01:45Z" level=error msg="jobs.batch \"cluster1-rmdata-xers\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.6.0 "timeoutSeconds": 10 "topologyKey": "kubernetes.io/hostname" "type": "ClusterIP", "type": "RollingUpdate", "value": "" "value": "10000" "value": "5432" "value": "5432" "value": "/backrestrepo/cluster1-backrest-shared-repo" "value": "/backrestrepo/cluster1-backrest-shared-repo" "value": "backup" "value": "cluster1" "value": "cluster1" "value": "cluster1-backrest-shared-repo" "value": "cluster1-backrest-shared-repo-57f668d478-qsc7n" "value": "cluster1-backrest-shared-repo-57f668d478-x4rln" "value": "cluster1-backrest-shared-repo-5f647d77b5-g5xnt" "value": "cluster1-backrest-shared-repo-5f647d77b5-t4ljf" "value": "cluster1-backrest-shared-repo-5f647d77b5-zsv7l" "value": "cluster1-backrest-shared-repo-65857c588c-fxb56" "value": "cluster1-backrest-shared-repo-65857c588c-pg2vr" "value": "cluster1-backrest-shared-repo-65857c588c-plqnc" "value": "cluster1-backrest-shared-repo-65857c588c-s6kmr" "value": "cluster1-backrest-shared-repo-85c794b5cb-gdbzd" "value": "cluster1-backrest-shared-repo-85c794b5cb-mcz8v" "value": "cluster1-backrest-shared-repo-85c794b5cb-nd2rj" "value": "cluster1-backrest-shared-repo-85c794b5cb-wbshm" "value": "cluster1-backrest-shared-repo-cbc58f67f-ndwjv" "value": "cluster1-repl1" "value": "cluster1-repl2" "value": "crunchy-pgha-scope" "value": "db" "value": "db" "value": " --db-host=10.44.218.12 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.18 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.24 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.29 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.34 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.40 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.45 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.50 --db-path=/pgdata/cluster1" "value": " --db-host=10.44.218.55 --db-path=/pgdata/cluster1" "value": "false" "value": "false" "valueFrom": { "value": "INFO" "value": "pgbackrest" "value": "pgbackrest-repo" "value": "/pgdata/cluster1" "value": "/pgdata/cluster1" "value": "/pgdata/cluster1-repl1" "value": "/pgdata/cluster1-repl1" "value": "/pgdata/cluster1-repl2" "value": "/pgdata/cluster1-repl2" "value": "posix" "value": "posix" "value": "postgres" "values": [ "value": "stanza-create" "value": "/tmp" "value": "/tmp" "value": "/tmp/nss_wrapper/pgbackrest-repo/group" "value": "/tmp/nss_wrapper/pgbackrest-repo/passwd" "value": "/tmp/nss_wrapper/postgres/group" "value": "/tmp/nss_wrapper/postgres/passwd" "value": "true" "value": "true" "value": "--type=full --db-host=10.44.216.20 --db-path=/pgdata/cluster1-repl1" "value": "--type=full --db-host=10.44.216.42 --db-path=/pgdata/cluster1-repl1" "value": "--type=full --db-host=10.44.217.25 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.217.47 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.217.56 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.217.64 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.12 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.18 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.24 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.26 --db-path=/pgdata/cluster1-repl2" "value": "--type=full --db-host=10.44.218.29 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.34 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.40 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.45 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.47 --db-path=/pgdata/cluster1-repl2" "value": "--type=full --db-host=10.44.218.50 --db-path=/pgdata/cluster1" "value": "--type=full --db-host=10.44.218.55 --db-path=/pgdata/cluster1" "value": "/usr/lib64/libnss_wrapper.so" "value": "/usr/lib64/libnss_wrapper.so" "value": "{vendor: \"crunchydata\"}" "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "volumeMounts": [ "volumeMounts": [{ "volumeMounts": [ "volumes": [ "volumes": [{ "weight": 1, + destroy_operator + kubectl_bin delete -f /tmp/tmp.09n2hxddFJ/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.ESivYcNvYF ++ mktemp + local LAST_ERR=/tmp/tmp.cZUSxzSsDq + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete -f /tmp/tmp.09n2hxddFJ/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete -f /tmp/tmp.09n2hxddFJ/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete -f /tmp/tmp.09n2hxddFJ/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.ESivYcNvYF + cat /tmp/tmp.cZUSxzSsDq Error from server (NotFound): error when deleting "/tmp/tmp.09n2hxddFJ/operator.yaml": serviceaccounts "pgo-deployer-sa" not found Error from server (NotFound): error when deleting "/tmp/tmp.09n2hxddFJ/operator.yaml": clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): error when deleting "/tmp/tmp.09n2hxddFJ/operator.yaml": configmaps "pgo-deployer-cm" not found Error from server (NotFound): error when deleting "/tmp/tmp.09n2hxddFJ/operator.yaml": clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-crb" not found Error from server (NotFound): error when deleting "/tmp/tmp.09n2hxddFJ/operator.yaml": jobs.batch "pgo-deploy" not found + rm /tmp/tmp.ESivYcNvYF /tmp/tmp.cZUSxzSsDq + return 1 + true + yq w -d4 /tmp/tmp.09n2hxddFJ/operator.yaml 'spec.template.spec.containers[0].env[0].value' uninstall + kubectl_bin apply -f /tmp/tmp.09n2hxddFJ/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.iZpmRXtEL5 ++ mktemp + local LAST_ERR=/tmp/tmp.uPSejMXlXr + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl apply -f /tmp/tmp.09n2hxddFJ/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.iZpmRXtEL5 serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb created job.batch/pgo-deploy created + cat /tmp/tmp.uPSejMXlXr + rm /tmp/tmp.iZpmRXtEL5 /tmp/tmp.uPSejMXlXr + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=version-service-20560 + retry=0 + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n version-service-20560 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.09n2hxddFJ/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.Pks8z5RJ2j ++ mktemp + local LAST_ERR=/tmp/tmp.GnoQqfoLEL + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete -f /tmp/tmp.09n2hxddFJ/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Pks8z5RJ2j serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.GnoQqfoLEL + rm /tmp/tmp.Pks8z5RJ2j /tmp/tmp.GnoQqfoLEL + return 0 + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + awk '{print$1}' ++ mktemp + xargs kubectl delete ns --grace-period=0 --force=true + local LAST_OUT=/tmp/tmp.P4tPYr8b0b ++ mktemp + local LAST_ERR=/tmp/tmp.NdSOHis5JA + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.P4tPYr8b0b + cat /tmp/tmp.NdSOHis5JA + rm /tmp/tmp.P4tPYr8b0b /tmp/tmp.NdSOHis5JA + return 0 Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. namespace "version-service-20560" force deleted + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.wFLQDoKkvG ++ mktemp + local LAST_ERR=/tmp/tmp.CnqZAhxG0H + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.wFLQDoKkvG + cat /tmp/tmp.CnqZAhxG0H Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.wFLQDoKkvG /tmp/tmp.CnqZAhxG0H + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.VLHAKIuplJ ++ mktemp + local LAST_ERR=/tmp/tmp.VlLQfA7nGY + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-723-7ed19e708-2-version-service + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.VLHAKIuplJ + cat /tmp/tmp.VlLQfA7nGY Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.VLHAKIuplJ /tmp/tmp.VlLQfA7nGY + return 1 + true + rm -rf /tmp/tmp.09n2hxddFJ