++ echo 'Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/logs/affinity.log' Log: /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/logs/affinity.log ++ '[' -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP= ++ oc get projects error: the server doesn't have a resource type "projects" +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep '\-eks\-' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ '[' ']' +++ kubectl version -o json +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ /usr/bin/sed -r 's/[^0-9.]+//g' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ KUBE_VERSION=1.26 +++ helm version -c +++ /usr/bin/sed -re 's/.*SemVer:"([^"]+)".*/\1/; s/.*\bVersion:"([^"]+)".*/\1/' ++ HELM_VERSION=v3.15.1 ++ '[' v3 == v2 ']' + create_namespace affinity-12045 + local namespace=affinity-12045 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + awk '{print$1}' + cleanup_rbac + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.rJ9FjuKsGj + local LAST_OUT=/tmp/tmp.FZGTAkXyVf ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.7SFyvSOG0t + local exit_status=0 + local timeout=4 + set +o errexit + local LAST_ERR=/tmp/tmp.K60gVFCIef + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FZGTAkXyVf + cat /tmp/tmp.K60gVFCIef + rm /tmp/tmp.FZGTAkXyVf /tmp/tmp.K60gVFCIef + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.rJ9FjuKsGj + cat /tmp/tmp.7SFyvSOG0t Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.rJ9FjuKsGj /tmp/tmp.7SFyvSOG0t + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level ++ mktemp + local LAST_OUT=/tmp/tmp.3ei2RzH8nB ++ mktemp + local LAST_ERR=/tmp/tmp.e6pOMSBBXQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr chaos-mesh-chaos-controller-manager-target-namespace chaos-mesh-chaos-controller-manager-cluster-level + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.3ei2RzH8nB + cat /tmp/tmp.e6pOMSBBXQ Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + rm /tmp/tmp.3ei2RzH8nB /tmp/tmp.e6pOMSBBXQ + return 1 + true + kubectl delete MutatingWebhookConfiguration/chaos-mesh-mutation ValidatingWebhookConfiguration/chaos-mesh-validation ValidatingWebhookConfiguration/validate-auth Error from server (NotFound): mutatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-mutation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-validation" not found Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "validate-auth" not found + true + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin delete namespace affinity-12045 ++ mktemp + local LAST_OUT=/tmp/tmp.Ttcz81llYl ++ mktemp + local LAST_ERR=/tmp/tmp.B9ImVmvmQW + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete namespace affinity-12045 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete namespace affinity-12045 + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete namespace affinity-12045 + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.Ttcz81llYl + cat /tmp/tmp.B9ImVmvmQW Error from server (NotFound): namespaces "affinity-12045" not found + rm /tmp/tmp.Ttcz81llYl /tmp/tmp.B9ImVmvmQW + return 1 + : + wait_for_delete namespace/affinity-12045 + local res=namespace/affinity-12045 + set +o xtrace namespace/affinity-12045 - Error from server (NotFound): namespaces "affinity-12045" not found + kubectl_bin create namespace affinity-12045 ++ mktemp + local LAST_OUT=/tmp/tmp.glmX3L6sRI ++ mktemp + local LAST_ERR=/tmp/tmp.Cduj5v4N2Y + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl create namespace affinity-12045 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.glmX3L6sRI namespace/affinity-12045 created + cat /tmp/tmp.Cduj5v4N2Y + rm /tmp/tmp.glmX3L6sRI /tmp/tmp.Cduj5v4N2Y + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.D047U3dAWa +++ mktemp ++ local LAST_ERR=/tmp/tmp.z3KkkeKFK2 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.D047U3dAWa ++ cat /tmp/tmp.z3KkkeKFK2 ++ rm /tmp/tmp.D047U3dAWa /tmp/tmp.z3KkkeKFK2 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-sandbox --namespace=affinity-12045 ++ mktemp + local LAST_OUT=/tmp/tmp.PTz7BXeWxM ++ mktemp + local LAST_ERR=/tmp/tmp.BzjZFtFdtp + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl config set-context gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-sandbox --namespace=affinity-12045 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.PTz7BXeWxM Context "gke_cloud-dev-112233_us-central1-c_jen-pg-776-0087ef129-2-sandbox" modified. + cat /tmp/tmp.BzjZFtFdtp + rm /tmp/tmp.PTz7BXeWxM /tmp/tmp.BzjZFtFdtp + return 0 + deploy_operator + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml + local namespace_mode=disabled + local action=install + local namespace=affinity-12045 + local dont_send_telemetry=false + local version_service_uri_override= + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + prepare_operator_yaml '' /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml disabled install affinity-12045 false + local pull_secret_name= + local operator_manifest=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml + local namespace_mode=disabled + local operator_action=install + local namespace=affinity-12045 + local dont_send_telemetry=false + yq r -d2 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml 'data[values.yaml]' + /usr/bin/sed -e 's#pgo_operator_namespace: .*#pgo_operator_namespace: "affinity-12045"#g' + /usr/bin/sed -e 's#namespace_mode: .*#namespace_mode: "disabled"#g' + /usr/bin/sed -e 's#^disable_telemetry: .*#disable_telemetry: "false"#g' + cat /dev/fd/63 - + /usr/bin/sed -e 's#^namespace: .*#namespace: "affinity-12045"#g' ++ echo perconalab/percona-postgresql-operator:PR-776-0087ef129 ++ echo 'crunchy_debug: "true"' ++ echo perconalab/percona-postgresql-operator:PR-776-0087ef129 ++ cut -d: -f2 ++ cut -d: -f1 + /usr/bin/sed -e 's#pgo_image_tag: .*#pgo_image_tag: "PR-776-0087ef129"#g' + /usr/bin/sed -e 's#pgo_image_prefix: .*#pgo_image_prefix: "perconalab/percona-postgresql-operator"#g' + [[ -n '' ]] + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + yq w '-d*' /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/operator.yaml metadata.namespace affinity-12045 + yq w -d4 - 'spec.template.spec.containers[0].image' perconalab/percona-postgresql-operator:PR-776-0087ef129-pgo-deployer + yq w -d4 - 'spec.template.spec.containers[0].env[0].value' install + yq w -d3 - 'subjects[0].namespace' affinity-12045 ++ cat /tmp/tmp.Hi0Eb7SlVS/operator.ini + yq w -d2 - -d2 'data[values.yaml]' 'crunchy_debug: "true" archive_mode: "true" archive_timeout: "60" ccp_image_pull_secret: "" ccp_image_pull_secret_manifest: "" create_rbac: "true" delete_operator_namespace: "false" delete_watched_namespaces: "false" disable_telemetry: "false" namespace: "affinity-12045" namespace_mode: "disabled" pgo_image_prefix: "perconalab/percona-postgresql-operator" pgo_image_pull_policy: "Always" pgo_image_pull_secret: "" pgo_image_pull_secret_manifest: "" pgo_image_tag: "PR-776-0087ef129" pgo_operator_namespace: "affinity-12045"' + [[ -n '' ]] + kubectl_bin apply -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.6crW2W8RWJ ++ mktemp + local LAST_ERR=/tmp/tmp.xrPj00PavC + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.6crW2W8RWJ serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb created job.batch/pgo-deploy created + cat /tmp/tmp.xrPj00PavC + rm /tmp/tmp.6crW2W8RWJ /tmp/tmp.xrPj00PavC + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 15 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 16 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 17 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 18 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 19 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 20 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 21 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 22 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 23 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 24 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 25 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 26 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 27 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 28 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 29 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 30 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 31 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 32 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 33 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 34 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.l8LHq9ZIwo ++ mktemp + local LAST_ERR=/tmp/tmp.GPTHrZtzGx + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.l8LHq9ZIwo serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.GPTHrZtzGx + rm /tmp/tmp.l8LHq9ZIwo /tmp/tmp.GPTHrZtzGx + return 0 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nm7rLpcxPy +++ mktemp ++ local LAST_ERR=/tmp/tmp.f8D6P6HDqQ ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.nm7rLpcxPy ++ cat /tmp/tmp.f8D6P6HDqQ ++ rm /tmp/tmp.nm7rLpcxPy /tmp/tmp.f8D6P6HDqQ ++ return 0 + wait_pod postgres-operator-775b589b68-gf5sx + local pod=postgres-operator-775b589b68-gf5sx + local ns= + set +o xtrace postgres-operator-775b589b68-gf5sxtrue + [[ -n '' ]] + cluster=some-name + spinup_pgcluster some-name /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=affinity-12045 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets some-name affinity-12045 + local cluster=some-name + local namespace=affinity-12045 + yq w - metadata.name some-name-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n affinity-12045 apply -f - + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.1BQfOqJSOB ++ mktemp + local LAST_ERR=/tmp/tmp.vN5wu4QS0e + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.1BQfOqJSOB secret/some-name-users-secret created + cat /tmp/tmp.vN5wu4QS0e + rm /tmp/tmp.1BQfOqJSOB /tmp/tmp.vN5wu4QS0e + return 0 + yq w - metadata.name some-name-pmm-secret + kubectl_bin -n affinity-12045 apply -f - + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.cNBW9256CP ++ mktemp + local LAST_ERR=/tmp/tmp.y9quD9FmV5 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.cNBW9256CP secret/some-name-pmm-secret created + cat /tmp/tmp.y9quD9FmV5 + rm /tmp/tmp.cNBW9256CP /tmp/tmp.y9quD9FmV5 + return 0 + kubectl_bin -n affinity-12045 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.fL7ZaoG4k6 ++ mktemp + local LAST_ERR=/tmp/tmp.q2LC2st7FU + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.fL7ZaoG4k6 deployment.apps/pg-client created + cat /tmp/tmp.q2LC2st7FU + rm /tmp/tmp.fL7ZaoG4k6 /tmp/tmp.q2LC2st7FU + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml some-name false false false false false false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local name=some-name + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=false + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml metadata.annotations.current-primary some-name + yq w - metadata.labels.crunchy-pgha-scope some-name + yq w - metadata.labels.pg-cluster some-name + yq w - metadata.name some-name + yq w - spec.clustername some-name + yq w - spec.database some-name + yq w - metadata.labels.name some-name + yq w - metadata.labels.deployment-name some-name + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.name some-name + yq w - spec.user some-name + yq w - spec.keepData false + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.keepBackups false + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.secretsName some-name-users-secret + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + case ${podAntiAffinity} in + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.Hi0Eb7SlVS/cr.yaml -n affinity-12045 ++ mktemp + local LAST_OUT=/tmp/tmp.mfsapIcDvk ++ mktemp + local LAST_ERR=/tmp/tmp.N8VSJP51qH + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.Hi0Eb7SlVS/cr.yaml -n affinity-12045 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.mfsapIcDvk perconapgcluster.pg.percona.com/some-name created + cat /tmp/tmp.N8VSJP51qH + rm /tmp/tmp.mfsapIcDvk /tmp/tmp.N8VSJP51qH + return 0 + wait_deployment some-name-backrest-shared-repo + local name=some-name-backrest-shared-repo + local target_namespace=affinity-12045 + sleep 10 + set +o xtrace some-name-backrest-shared-repo.. + wait_deployment some-name + local name=some-name + local target_namespace=affinity-12045 + sleep 10 + set +o xtrace some-name......... + [[ false == \g\c\s\+* ]] + wait_job_completion some-name-stanza-create false + local job=some-name-stanza-create + local mandatory=false + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' some-name-stanza-create+ retry=0 ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-some-name + local job=backrest-backup-some-name + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' backrest-backup-some-name+ retry=0 ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 15 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency some-name + cluster_name=some-name + wait_cluster_status some-name 'pgcluster Initialized' + local cluster_name=some-name + local 'target_status=pgcluster Initialized' + local target_namespace=affinity-12045 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KjKrU8BZHf +++ mktemp ++ local LAST_ERR=/tmp/tmp.hOyuVIJ5ad ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.KjKrU8BZHf ++ cat /tmp/tmp.hOyuVIJ5ad ++ rm /tmp/tmp.KjKrU8BZHf /tmp/tmp.hOyuVIJ5ad ++ return 0 + status_query='kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.YmZHfdhwgq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.2HACOhQ7Xo +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.YmZHfdhwgq +++ cat /tmp/tmp.2HACOhQ7Xo +++ rm /tmp/tmp.YmZHfdhwgq /tmp/tmp.2HACOhQ7Xo +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.riizFqB3rT ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Xqc3ajHtmT +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.riizFqB3rT +++ cat /tmp/tmp.Xqc3ajHtmT +++ rm /tmp/tmp.riizFqB3rT /tmp/tmp.Xqc3ajHtmT +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OAwO6X3eiQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.izudytWCAf +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.OAwO6X3eiQ +++ cat /tmp/tmp.izudytWCAf +++ rm /tmp/tmp.OAwO6X3eiQ /tmp/tmp.izudytWCAf +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0CWN72dcWU ++++ mktemp +++ local LAST_ERR=/tmp/tmp.dURdhAKjAF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.0CWN72dcWU +++ cat /tmp/tmp.dURdhAKjAF +++ rm /tmp/tmp.0CWN72dcWU /tmp/tmp.dURdhAKjAF +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Fg9hylONCN ++++ mktemp +++ local LAST_ERR=/tmp/tmp.rhjnluDt3w +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Fg9hylONCN +++ cat /tmp/tmp.rhjnluDt3w +++ rm /tmp/tmp.Fg9hylONCN /tmp/tmp.rhjnluDt3w +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.s9GNwk5rg3 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.LfnpaaHFmD ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.s9GNwk5rg3 ++++ cat /tmp/tmp.LfnpaaHFmD ++++ rm /tmp/tmp.s9GNwk5rg3 /tmp/tmp.LfnpaaHFmD ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.x4Tp6v4Nsr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.o9NJLioyZ9 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.x4Tp6v4Nsr +++ cat /tmp/tmp.o9NJLioyZ9 +++ rm /tmp/tmp.x4Tp6v4Nsr /tmp/tmp.o9NJLioyZ9 +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@some-name.affinity-12045 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@some-name.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QEBZXL2IZ6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.AT7AjFeDqY ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.QEBZXL2IZ6 ++ cat /tmp/tmp.AT7AjFeDqY ++ rm /tmp/tmp.QEBZXL2IZ6 /tmp/tmp.AT7AjFeDqY ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.qLYk40cCkd ++ mktemp + local LAST_ERR=/tmp/tmp.o3BZ6YDSY0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.qLYk40cCkd + cat /tmp/tmp.o3BZ6YDSY0 + rm /tmp/tmp.qLYk40cCkd /tmp/tmp.o3BZ6YDSY0 + return 0 ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.zMBUt3mq3j ++++ mktemp +++ local LAST_ERR=/tmp/tmp.z9GkAbZiEI +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.zMBUt3mq3j +++ cat /tmp/tmp.z9GkAbZiEI +++ rm /tmp/tmp.zMBUt3mq3j /tmp/tmp.z9GkAbZiEI +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3dj2I1URml ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yf8QXNUfrV +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.3dj2I1URml +++ cat /tmp/tmp.yf8QXNUfrV +++ rm /tmp/tmp.3dj2I1URml /tmp/tmp.yf8QXNUfrV +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.cj3tAUf1Uq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LBp8aEjeGd +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.cj3tAUf1Uq +++ cat /tmp/tmp.LBp8aEjeGd +++ rm /tmp/tmp.cj3tAUf1Uq /tmp/tmp.LBp8aEjeGd +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.jyTsOPaz2W +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.YDSccezYkj ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.jyTsOPaz2W ++++ cat /tmp/tmp.YDSccezYkj ++++ rm /tmp/tmp.jyTsOPaz2W /tmp/tmp.YDSccezYkj ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Dg4vfhbSpF ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Xiy2T3CGm5 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Dg4vfhbSpF +++ cat /tmp/tmp.Xiy2T3CGm5 +++ rm /tmp/tmp.Dg4vfhbSpF /tmp/tmp.Xiy2T3CGm5 +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@some-name.affinity-12045 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@some-name.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hGRcSt9h3V +++ mktemp ++ local LAST_ERR=/tmp/tmp.CTA84a4Fln ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.hGRcSt9h3V ++ cat /tmp/tmp.CTA84a4Fln ++ rm /tmp/tmp.hGRcSt9h3V /tmp/tmp.CTA84a4Fln ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.09Jf0o3Zkj ++ mktemp + local LAST_ERR=/tmp/tmp.UwYwnDiQpV + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.09Jf0o3Zkj + cat /tmp/tmp.UwYwnDiQpV + rm /tmp/tmp.09Jf0o3Zkj /tmp/tmp.UwYwnDiQpV + return 0 ++ get_psql_user_pass postgres some-name ++ local user=postgres ++ local cluster=some-name ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Zoo0vIV0fp ++++ mktemp +++ local LAST_ERR=/tmp/tmp.55Z7EMlLur +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.Zoo0vIV0fp +++ cat /tmp/tmp.55Z7EMlLur +++ rm /tmp/tmp.Zoo0vIV0fp /tmp/tmp.55Z7EMlLur +++ return 0 ++ [[ xsome-name == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ceS969pBgx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.k5HfwVBFdu +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ceS969pBgx +++ cat /tmp/tmp.k5HfwVBFdu +++ rm /tmp/tmp.ceS969pBgx /tmp/tmp.k5HfwVBFdu +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.yN5hTCzAO8 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9YL880kCap +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.yN5hTCzAO8 +++ cat /tmp/tmp.9YL880kCap +++ rm /tmp/tmp.yN5hTCzAO8 /tmp/tmp.9YL880kCap +++ return 0 ++ [[ xsome-name-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.H0RMXL6H80 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.cmQcSTS2Z4 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.H0RMXL6H80 ++++ cat /tmp/tmp.cmQcSTS2Z4 ++++ rm /tmp/tmp.H0RMXL6H80 /tmp/tmp.cmQcSTS2Z4 ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6KCJbg3Ped ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vGmnqWm7dp +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.6KCJbg3Ped +++ cat /tmp/tmp.vGmnqWm7dp +++ rm /tmp/tmp.6KCJbg3Ped /tmp/tmp.vGmnqWm7dp +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@some-name.affinity-12045 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@some-name.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8Z6g8ZtQNP +++ mktemp ++ local LAST_ERR=/tmp/tmp.HzrPcVyml3 ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.8Z6g8ZtQNP ++ cat /tmp/tmp.HzrPcVyml3 ++ rm /tmp/tmp.8Z6g8ZtQNP /tmp/tmp.HzrPcVyml3 ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.Fa74xn7Kdn ++ mktemp + local LAST_ERR=/tmp/tmp.DVMTCskUx4 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Fa74xn7Kdn + cat /tmp/tmp.DVMTCskUx4 + rm /tmp/tmp.Fa74xn7Kdn /tmp/tmp.DVMTCskUx4 + return 0 + sleep 10 + compare_kubectl deployment/some-name + local resource=deployment/some-name + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/deployment_some-name.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name + yq d - metadata.managedFields + yq d - '**.creationTimestamp' + yq d - '**.namespace' ++ mktemp + yq d - metadata.deletionTimestamp + yq d - '**.uid' + local LAST_OUT=/tmp/tmp.FQ7YdHAnu6 + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - metadata.selfLink + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - '**.creationTimestamp' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**.image' ++ mktemp + yq d - '**."volume.kubernetes.io/storage-provisioner"' + local LAST_ERR=/tmp/tmp.W4CAB0cOyY + local exit_status=0 + local timeout=4 + set +o errexit + yq d - '**."kubernetes.io/pvc-protection"' + yq d - spec.volumeMode + yq d - '**.storageClassName' + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**.volumeName' + yq d - '**.finalizers' + yq d - '**."percona.com/*"' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' ++ seq 0 2 + yq d - '**.healthCheckNodePort' + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name + yq d - spec.nodeName + yq d - '**.imagePullSecrets' + yq d - '**.internalTrafficPolicy' + yq d - '**.procMount' + yq d - '**.ipFamilies' + yq d - '**.enableServiceLinks' + yq d - spec.podReplacementPolicy + yq d - spec.manualSelector + yq d - '**.(name==NAMESPACE)' + yq d - '**.clusterIPs' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.dataSource' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.controller-uid' + yq d - '**.nodePort' + yq d - '**.preemptionPolicy' + yq d - status + yq d - '**.ipFamilyPolicy' + yq d - '**.(name==suffix)' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.(name==PODNAME)' + yq d - '**.clusterIP' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - metadata.resourceVersion + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.FQ7YdHAnu6 + cat /tmp/tmp.W4CAB0cOyY + rm /tmp/tmp.FQ7YdHAnu6 /tmp/tmp.W4CAB0cOyY + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/deployment_some-name.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name.yml /tmp/tmp.Hi0Eb7SlVS/deployment_some-name.yml + compare_kubectl deployment/some-name-backrest-shared-repo + local resource=deployment/some-name-backrest-shared-repo + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-backrest-shared-repo.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/deployment_some-name-backrest-shared-repo.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-backrest-shared-repo + yq d - metadata.managedFields + yq d - '**.creationTimestamp' + yq d - '**.uid' + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - '**.namespace' + yq d - metadata.selfLink + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - metadata.deletionTimestamp + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - '**.image' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**.clusterIP' + yq d - '**.clusterIPs' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - '**.dataSource' + yq d - '**.procMount' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**.storageClassName' + yq d - '**.creationTimestamp' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**.finalizers' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - spec.volumeMode + yq d - '**.volumeName' + yq d - spec.manualSelector + yq d - spec.nodeName + yq d - '**."percona.com/*"' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**.(name==suffix)' + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - spec.podReplacementPolicy + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - status + yq d - '**.(name==NAMESPACE)' + yq d - '**.healthCheckNodePort' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.nodePort' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.imagePullSecrets' + yq d - '**.enableServiceLinks' + yq d - '**.(name==COMMAND_OPTS)' + yq d - metadata.resourceVersion + yq d - '**.(name==PODNAME)' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.controller-uid' + yq d - '**.preemptionPolicy' + yq d - '**.ipFamilies' + yq d - '**.ipFamilyPolicy' + yq d - '**.internalTrafficPolicy' ++ mktemp + local LAST_OUT=/tmp/tmp.R4AfSC8ntE ++ mktemp + local LAST_ERR=/tmp/tmp.pHgPJsDQOJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-backrest-shared-repo + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.R4AfSC8ntE + cat /tmp/tmp.pHgPJsDQOJ + rm /tmp/tmp.R4AfSC8ntE /tmp/tmp.pHgPJsDQOJ + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-backrest-shared-repo.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-backrest-shared-repo.yml /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-backrest-shared-repo.yml + compare_kubectl job/backrest-backup-some-name + local resource=job/backrest-backup-some-name + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/job_backrest-backup-some-name.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + yq d - metadata.managedFields + yq d - '**.creationTimestamp' + kubectl_bin get -o yaml job/backrest-backup-some-name ++ mktemp + yq d - metadata.selfLink + yq d - metadata.deletionTimestamp + local LAST_OUT=/tmp/tmp.ztSWS5lxjE + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - metadata.resourceVersion + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' ++ mktemp + yq d - '**.image' + yq d - '**.clusterIP' + yq d - '**.storageClassName' + yq d - '**.finalizers' + local LAST_ERR=/tmp/tmp.Bp6wSKi74s + local exit_status=0 + local timeout=4 + set +o errexit + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**.namespace' + yq d - '**.uid' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - '**.volumeName' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**.clusterIPs' + yq d - '**.creationTimestamp' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**.procMount' + yq d - '**.dataSource' + yq d - spec.volumeMode ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml job/backrest-backup-some-name + yq d - spec.podReplacementPolicy + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - spec.manualSelector + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - spec.nodeName + yq d - '**."percona.com/*"' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.imagePullSecrets' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.healthCheckNodePort' + yq d - '**.enableServiceLinks' + yq d - '**.nodePort' + yq d - '**.(name==COMMAND_OPTS)' + yq d - status + yq d - '**.(name==PODNAME)' + yq d - '**.(name==suffix)' + yq d - '**.(name==NAMESPACE)' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.controller-uid' + yq d - '**.ipFamilies' + yq d - '**.preemptionPolicy' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**.ipFamilyPolicy' + yq d - '**.internalTrafficPolicy' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ztSWS5lxjE + cat /tmp/tmp.Bp6wSKi74s + rm /tmp/tmp.ztSWS5lxjE /tmp/tmp.Bp6wSKi74s + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml kind ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml metadata.annotations + [[ x{} == \x ]] ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml metadata.annotations + [[ x{} == \x\{\} ]] + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml metadata.annotations + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml metadata.generation + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml spec.completionMode + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml spec.suspend + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/job_backrest-backup-some-name.yml /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name.yml + create_backup some-name preferred + local cluster=some-name + local bckp_prefix=preferred + local bckp_type=full + yq w - metadata.name preferred-some-name + yq w - spec.name preferred-some-name + yq w - spec.namespace affinity-12045 + yq w - spec.parameters.job-name preferred-some-name + kubectl_bin apply -f - + yq w - spec.parameters.pg-cluster some-name + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster some-name + yq w --style=single -- - spec.parameters.backrest-opts --type=full ++ mktemp ++ kubectl_bin get pods --selector=name=some-name-backrest-shared-repo,pg-cluster=some-name -o 'jsonpath={.items[].metadata.name}' +++ mktemp + local LAST_OUT=/tmp/tmp.J1YT6LADuL ++ mktemp + local LAST_ERR=/tmp/tmp.UIHoxu86Uy + local exit_status=0 + local timeout=4 + set +o errexit ++ local LAST_OUT=/tmp/tmp.NLLcJu7hwA ++ seq 0 2 +++ mktemp + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f - ++ local LAST_ERR=/tmp/tmp.DSoj6R9VMe ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=some-name-backrest-shared-repo,pg-cluster=some-name -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.NLLcJu7hwA ++ cat /tmp/tmp.DSoj6R9VMe ++ rm /tmp/tmp.NLLcJu7hwA /tmp/tmp.DSoj6R9VMe ++ return 0 + yq w - spec.parameters.podname some-name-backrest-shared-repo-6f8f947dbc-96d8f + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.J1YT6LADuL pgtask.pg.percona.com/preferred-some-name created + cat /tmp/tmp.UIHoxu86Uy + rm /tmp/tmp.J1YT6LADuL /tmp/tmp.UIHoxu86Uy + return 0 + sleep 10 + wait_job_completion preferred-some-name + local job=preferred-some-name + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' preferred-some-name+ retry=0 ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/preferred-some-name -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete perconapgcluster some-name ++ mktemp + local LAST_OUT=/tmp/tmp.GteGrf2kC2 ++ mktemp + local LAST_ERR=/tmp/tmp.l3oyHYNgwF + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete perconapgcluster some-name + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.GteGrf2kC2 perconapgcluster.pg.percona.com "some-name" deleted + cat /tmp/tmp.l3oyHYNgwF + rm /tmp/tmp.GteGrf2kC2 /tmp/tmp.l3oyHYNgwF + return 0 + sleep 30 + cluster=some-name-required + spinup_pgcluster some-name-required /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml '' '' required + local cluster=some-name-required + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=required + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=affinity-12045 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets some-name-required affinity-12045 + local cluster=some-name-required + local namespace=affinity-12045 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name some-name-required-users-secret + yq w - stringData.some-name some-name_pass + kubectl_bin -n affinity-12045 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.MnCWhBYA7k ++ mktemp + local LAST_ERR=/tmp/tmp.uaGKyPPpXM + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.MnCWhBYA7k secret/some-name-required-users-secret created + cat /tmp/tmp.uaGKyPPpXM + rm /tmp/tmp.MnCWhBYA7k /tmp/tmp.uaGKyPPpXM + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name some-name-required-pmm-secret + kubectl_bin -n affinity-12045 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.pUpVroTWr7 ++ mktemp + local LAST_ERR=/tmp/tmp.9hMZX7sKRs + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.pUpVroTWr7 secret/some-name-required-pmm-secret created + cat /tmp/tmp.9hMZX7sKRs + rm /tmp/tmp.pUpVroTWr7 /tmp/tmp.9hMZX7sKRs + return 0 + kubectl_bin -n affinity-12045 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.IC7RCyjvwb ++ mktemp + local LAST_ERR=/tmp/tmp.kXeJofNzYI + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.IC7RCyjvwb deployment.apps/pg-client unchanged + cat /tmp/tmp.kXeJofNzYI + rm /tmp/tmp.IC7RCyjvwb /tmp/tmp.kXeJofNzYI + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml some-name-required false false false false required false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local name=some-name-required + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=required + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml metadata.annotations.current-primary some-name-required + yq w - metadata.labels.deployment-name some-name-required + yq w - metadata.labels.pg-cluster some-name-required + yq w - metadata.labels.name some-name-required + yq w - spec.database some-name-required + yq w - spec.keepBackups false + yq w - spec.keepData false + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.secretsName some-name-required-users-secret + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.clustername some-name-required + yq w - metadata.name some-name-required + yq w - metadata.labels.crunchy-pgha-scope some-name-required + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.name some-name-required + yq w - spec.user some-name + case ${podAntiAffinity} in + yq w - spec.backup.affinity.antiAffinityType required + yq w - spec.pgBouncer.affinity.antiAffinityType required + yq w /tmp/tmp.Hi0Eb7SlVS/cr.yaml spec.pgPrimary.affinity.antiAffinityType required + mv /tmp/tmp.Hi0Eb7SlVS/cr.podAffinity.yaml /tmp/tmp.Hi0Eb7SlVS/cr.yaml + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.Hi0Eb7SlVS/cr.yaml -n affinity-12045 ++ mktemp + local LAST_OUT=/tmp/tmp.zKBo9WFc5Y ++ mktemp + local LAST_ERR=/tmp/tmp.mUmBNWLCLt + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.Hi0Eb7SlVS/cr.yaml -n affinity-12045 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.zKBo9WFc5Y perconapgcluster.pg.percona.com/some-name-required created + cat /tmp/tmp.mUmBNWLCLt + rm /tmp/tmp.zKBo9WFc5Y /tmp/tmp.mUmBNWLCLt + return 0 + wait_deployment some-name-required-backrest-shared-repo + local name=some-name-required-backrest-shared-repo + local target_namespace=affinity-12045 + sleep 10 + set +o xtrace some-name-required-backrest-shared-repo.... + wait_deployment some-name-required + local name=some-name-required + local target_namespace=affinity-12045 + sleep 10 + set +o xtrace some-name-required......... + [[ false == \g\c\s\+* ]] + wait_job_completion some-name-required-stanza-create false + local job=some-name-required-stanza-create + local mandatory=false + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/some-name-required-stanza-create -o 'jsonpath={.metadata.name}' some-name-required-stanza-create+ retry=0 ++ kubectl_bin -n affinity-12045 get job/some-name-required-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-required-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-required-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-required-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-required-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-required-stanza-create -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_job_completion backrest-backup-some-name-required + local job=backrest-backup-some-name-required + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' + sleep 1 + echo -n . .+ let retry+=1 + [[ 1 -ge 120 ]] + [[ 1 -ge 120 ]] + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' backrest-backup-some-name-required+ retry=0 ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-required -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency some-name-required + cluster_name=some-name-required + wait_cluster_status some-name-required 'pgcluster Initialized' + local cluster_name=some-name-required + local 'target_status=pgcluster Initialized' + local target_namespace=affinity-12045 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bxYL6DU0iF +++ mktemp ++ local LAST_ERR=/tmp/tmp.IxaAH612bU ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.bxYL6DU0iF ++ cat /tmp/tmp.IxaAH612bU ++ rm /tmp/tmp.bxYL6DU0iF /tmp/tmp.IxaAH612bU ++ return 0 + status_query='kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.gXAW2fhxYn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bidgPztJXd +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.gXAW2fhxYn +++ cat /tmp/tmp.bidgPztJXd +++ rm /tmp/tmp.gXAW2fhxYn /tmp/tmp.bidgPztJXd +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.63iFG4HK0m ++++ mktemp +++ local LAST_ERR=/tmp/tmp.IWQXml9JsF +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.63iFG4HK0m +++ cat /tmp/tmp.IWQXml9JsF +++ rm /tmp/tmp.63iFG4HK0m /tmp/tmp.IWQXml9JsF +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres some-name-required ++ local user=postgres ++ local cluster=some-name-required ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1EkGX5stni ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Hr4jU0LZnx +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.1EkGX5stni +++ cat /tmp/tmp.Hr4jU0LZnx +++ rm /tmp/tmp.1EkGX5stni /tmp/tmp.Hr4jU0LZnx +++ return 0 ++ [[ xsome-name-required == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OSeeGmiFoK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.thxTlYHzyz +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.OSeeGmiFoK +++ cat /tmp/tmp.thxTlYHzyz +++ rm /tmp/tmp.OSeeGmiFoK /tmp/tmp.thxTlYHzyz +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VL8pcWfP97 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LKdz0lT0jB +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.VL8pcWfP97 +++ cat /tmp/tmp.LKdz0lT0jB +++ rm /tmp/tmp.VL8pcWfP97 /tmp/tmp.LKdz0lT0jB +++ return 0 ++ [[ xsome-name-required-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.ZMM2RI0IjK +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.baaHSyuSB5 ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.ZMM2RI0IjK ++++ cat /tmp/tmp.baaHSyuSB5 ++++ rm /tmp/tmp.ZMM2RI0IjK /tmp/tmp.baaHSyuSB5 ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-required-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.2sVqyuYBfA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.RxaYsbdsHU +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-required-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.2sVqyuYBfA +++ cat /tmp/tmp.RxaYsbdsHU +++ rm /tmp/tmp.2sVqyuYBfA /tmp/tmp.RxaYsbdsHU +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@some-name-required.affinity-12045 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@some-name-required.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dc9kx0KKb7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Grl3RA0u2G ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.Dc9kx0KKb7 ++ cat /tmp/tmp.Grl3RA0u2G ++ rm /tmp/tmp.Dc9kx0KKb7 /tmp/tmp.Grl3RA0u2G ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-required.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.Z5uGLtAolB ++ mktemp + local LAST_ERR=/tmp/tmp.Zag2JTBHGl + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-required.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Z5uGLtAolB + cat /tmp/tmp.Zag2JTBHGl + rm /tmp/tmp.Z5uGLtAolB /tmp/tmp.Zag2JTBHGl + return 0 ++ get_psql_user_pass postgres some-name-required ++ local user=postgres ++ local cluster=some-name-required ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0NJrZA7AMr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZjtiskfD5H +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.0NJrZA7AMr +++ cat /tmp/tmp.ZjtiskfD5H +++ rm /tmp/tmp.0NJrZA7AMr /tmp/tmp.ZjtiskfD5H +++ return 0 ++ [[ xsome-name-required == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.GLbZ0idiQj ++++ mktemp +++ local LAST_ERR=/tmp/tmp.deXofv9MMf +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.GLbZ0idiQj +++ cat /tmp/tmp.deXofv9MMf +++ rm /tmp/tmp.GLbZ0idiQj /tmp/tmp.deXofv9MMf +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.iQOt2p1xsA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.iGFwB51IOA +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.iQOt2p1xsA +++ cat /tmp/tmp.iGFwB51IOA +++ rm /tmp/tmp.iQOt2p1xsA /tmp/tmp.iGFwB51IOA +++ return 0 ++ [[ xsome-name-required-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.J1qObl7Wwp +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.5dj0ud8B0s ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.J1qObl7Wwp ++++ cat /tmp/tmp.5dj0ud8B0s ++++ rm /tmp/tmp.J1qObl7Wwp /tmp/tmp.5dj0ud8B0s ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-required-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VmFGRYEICI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.K4Bz6oE2MJ +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-required-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.VmFGRYEICI +++ cat /tmp/tmp.K4Bz6oE2MJ +++ rm /tmp/tmp.VmFGRYEICI /tmp/tmp.K4Bz6oE2MJ +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@some-name-required.affinity-12045 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@some-name-required.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vufDnHdWNo +++ mktemp ++ local LAST_ERR=/tmp/tmp.LAdV2mR6Qb ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.vufDnHdWNo ++ cat /tmp/tmp.LAdV2mR6Qb ++ rm /tmp/tmp.vufDnHdWNo /tmp/tmp.LAdV2mR6Qb ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-required.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.v8BwIN0Yd2 ++ mktemp + local LAST_ERR=/tmp/tmp.gvbBCeRmJY + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-required.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.v8BwIN0Yd2 + cat /tmp/tmp.gvbBCeRmJY + rm /tmp/tmp.v8BwIN0Yd2 /tmp/tmp.gvbBCeRmJY + return 0 ++ get_psql_user_pass postgres some-name-required ++ local user=postgres ++ local cluster=some-name-required ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.xfGDXLJIWr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5WGQdvhiVe +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.xfGDXLJIWr +++ cat /tmp/tmp.5WGQdvhiVe +++ rm /tmp/tmp.xfGDXLJIWr /tmp/tmp.5WGQdvhiVe +++ return 0 ++ [[ xsome-name-required == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pvtjwnWwwb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.F0f9DuFK6f +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name-required -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.pvtjwnWwwb +++ cat /tmp/tmp.F0f9DuFK6f +++ rm /tmp/tmp.pvtjwnWwwb /tmp/tmp.F0f9DuFK6f +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ox3Vggwf7k ++++ mktemp +++ local LAST_ERR=/tmp/tmp.mOc122iP05 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.ox3Vggwf7k +++ cat /tmp/tmp.mOc122iP05 +++ rm /tmp/tmp.ox3Vggwf7k /tmp/tmp.mOc122iP05 +++ return 0 ++ [[ xsome-name-required-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.ATv5u67RLW +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.3ianm4xoXt ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name-required -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.ATv5u67RLW ++++ cat /tmp/tmp.3ianm4xoXt ++++ rm /tmp/tmp.ATv5u67RLW /tmp/tmp.3ianm4xoXt ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-required-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wXBKcH0y54 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.54rRTzMkfj +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-required-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.wXBKcH0y54 +++ cat /tmp/tmp.54rRTzMkfj +++ rm /tmp/tmp.wXBKcH0y54 /tmp/tmp.54rRTzMkfj +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@some-name-required.affinity-12045 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@some-name-required.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.236AXHtzCK +++ mktemp ++ local LAST_ERR=/tmp/tmp.KAd9KEPHSC ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.236AXHtzCK ++ cat /tmp/tmp.KAd9KEPHSC ++ rm /tmp/tmp.236AXHtzCK /tmp/tmp.KAd9KEPHSC ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-required.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.uWXNpN8zRe ++ mktemp + local LAST_ERR=/tmp/tmp.I6birN89aQ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-required.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.uWXNpN8zRe + cat /tmp/tmp.I6birN89aQ + rm /tmp/tmp.uWXNpN8zRe /tmp/tmp.I6birN89aQ + return 0 + sleep 10 + compare_kubectl deployment/some-name-required + local resource=deployment/some-name-required + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-required.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/deployment_some-name-required.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-required + yq d - metadata.managedFields + yq d - '**.creationTimestamp' ++ mktemp + yq d - '**.namespace' + local LAST_OUT=/tmp/tmp.MEQwvne1kq + yq d - metadata.deletionTimestamp + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**.creationTimestamp' ++ mktemp + local LAST_ERR=/tmp/tmp.DdkLHoHINy + local exit_status=0 + local timeout=4 + set +o errexit + yq d - '**.volumeName' + yq d - '**.procMount' + yq d - '**.finalizers' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - spec.volumeMode + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - spec.nodeName ++ seq 0 2 + yq d - '**."volume.kubernetes.io/selected-node"' + for i in '$(seq 0 2)' + yq d - '**.storageClassName' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-required + yq d - '**.image' + yq d - '**.(name==COMMAND_OPTS)' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - '**.healthCheckNodePort' + yq d - '**.imagePullSecrets' + yq d - '**.nodePort' + yq d - '**.ipFamilyPolicy' + yq d - status + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**.(name==suffix)' + yq d - '**.dataSource' + yq d - '**.(name==NAMESPACE)' + yq d - spec.podReplacementPolicy + yq d - spec.manualSelector + yq d - '**.(name==PODNAME)' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**."percona.com/*"' + yq d - '**.internalTrafficPolicy' + yq d - '**.enableServiceLinks' + yq d - '**.controller-uid' + yq d - '**.preemptionPolicy' + yq d - '**.ipFamilies' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - '**.clusterIPs' + yq d - '**.clusterIP' + yq d - metadata.selfLink + yq d - metadata.resourceVersion + yq d - '**.uid' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.MEQwvne1kq + cat /tmp/tmp.DdkLHoHINy + rm /tmp/tmp.MEQwvne1kq /tmp/tmp.DdkLHoHINy + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-required.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-required.yml /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-required.yml + compare_kubectl deployment/some-name-required-backrest-shared-repo + local resource=deployment/some-name-required-backrest-shared-repo + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-required-backrest-shared-repo.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/deployment_some-name-required-backrest-shared-repo.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-required-backrest-shared-repo + yq d - '**.creationTimestamp' + yq d - '**.namespace' + yq d - '**.uid' + yq d - metadata.resourceVersion + yq d - metadata.selfLink + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**.creationTimestamp' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - '**.image' + yq d - '**.procMount' + yq d - '**.clusterIP' ++ mktemp + yq d - '**.storageClassName' + local LAST_OUT=/tmp/tmp.YYuh5yG34Q + yq d - '**.finalizers' + yq d - '**.clusterIPs' + yq d - '**.dataSource' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**.volumeName' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - spec.nodeName + yq d - spec.volumeMode + yq d - '**.internalTrafficPolicy' + yq d - spec.podReplacementPolicy + yq d - spec.manualSelector + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**."percona.com/*"' + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**.ipFamilies' + yq d - '**.preemptionPolicy' + yq d - '**.nodePort' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - '**.ipFamilyPolicy' + yq d - '**.imagePullSecrets' + yq d - '**.healthCheckNodePort' + yq d - status + yq d - '**.enableServiceLinks' + yq d - '**.(name==suffix)' + yq d - '**.(name==NAMESPACE)' ++ mktemp + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.controller-uid' + yq d - '**.(name==PODNAME)' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + local LAST_ERR=/tmp/tmp.vpr2yndkGk + local exit_status=0 + local timeout=4 + set +o errexit + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - 'spec.volumeClaimTemplates.*.kind' ++ seq 0 2 + yq d - 'metadata.ownerReferences.*.apiVersion' + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-required-backrest-shared-repo + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - metadata.deletionTimestamp + yq d - metadata.managedFields + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.YYuh5yG34Q + cat /tmp/tmp.vpr2yndkGk + rm /tmp/tmp.YYuh5yG34Q /tmp/tmp.vpr2yndkGk + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-required-backrest-shared-repo.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-required-backrest-shared-repo.yml /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-required-backrest-shared-repo.yml + compare_kubectl job/backrest-backup-some-name-required + local resource=job/backrest-backup-some-name-required + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/job_backrest-backup-some-name-required.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + yq d - metadata.managedFields + yq d - '**.creationTimestamp' + kubectl_bin get -o yaml job/backrest-backup-some-name-required + yq d - metadata.selfLink + yq d - '**.uid' + yq d - metadata.resourceVersion + yq d - '**.namespace' ++ mktemp + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - metadata.deletionTimestamp + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - '**.creationTimestamp' + yq d - '**.image' + yq d - '**.clusterIPs' + yq d - '**.clusterIP' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - '**.healthCheckNodePort' + yq d - spec.manualSelector + yq d - '**.dataSource' + yq d - '**.(name==suffix)' + yq d - '**.nodePort' + local LAST_OUT=/tmp/tmp.cqJqDDxdVW + yq d - '**."percona.com/*"' + yq d - '**.(name==NAMESPACE)' + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.(name==PODNAME)' + yq d - '**.controller-uid' + yq d - '**.ipFamilyPolicy' + yq d - '**.imagePullSecrets' + yq d - '**.internalTrafficPolicy' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - spec.nodeName + yq d - '**.ipFamilies' + yq d - '**.enableServiceLinks' + yq d - status + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.finalizers' + yq d - '**.volumeName' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - spec.volumeMode + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.preemptionPolicy' + yq d - spec.podReplacementPolicy + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.storageClassName' ++ mktemp + yq d - '**.procMount' + local LAST_ERR=/tmp/tmp.Mf5KRsvD3D + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml job/backrest-backup-some-name-required + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.cqJqDDxdVW + cat /tmp/tmp.Mf5KRsvD3D + rm /tmp/tmp.cqJqDDxdVW /tmp/tmp.Mf5KRsvD3D + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml kind ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml metadata.annotations + [[ x{} == \x ]] ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml metadata.annotations + [[ x{} == \x\{\} ]] + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml metadata.annotations + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml metadata.generation + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml spec.completionMode + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml spec.suspend + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/job_backrest-backup-some-name-required.yml /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-required.yml + create_backup some-name-required required + local cluster=some-name-required + local bckp_prefix=required + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster some-name-required + yq w - metadata.name required-some-name-required + yq w - spec.name required-some-name-required + yq w --style=single -- - spec.parameters.backrest-opts --type=full + kubectl_bin apply -f - + yq w - spec.namespace affinity-12045 + yq w - spec.parameters.job-name required-some-name-required + yq w - spec.parameters.pg-cluster some-name-required ++ mktemp + local LAST_OUT=/tmp/tmp.kjir4rus8l ++ kubectl_bin get pods --selector=name=some-name-required-backrest-shared-repo,pg-cluster=some-name-required -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.pSQ2ERYAm0 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 ++ local LAST_OUT=/tmp/tmp.aN6SfR0Ftf + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f - +++ mktemp ++ local LAST_ERR=/tmp/tmp.L7DxZAtewx ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=some-name-required-backrest-shared-repo,pg-cluster=some-name-required -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.aN6SfR0Ftf ++ cat /tmp/tmp.L7DxZAtewx ++ rm /tmp/tmp.aN6SfR0Ftf /tmp/tmp.L7DxZAtewx ++ return 0 + yq w - spec.parameters.podname some-name-required-backrest-shared-repo-8558c9b6fc-2cdk9 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.kjir4rus8l pgtask.pg.percona.com/required-some-name-required created + cat /tmp/tmp.pSQ2ERYAm0 + rm /tmp/tmp.kjir4rus8l /tmp/tmp.pSQ2ERYAm0 + return 0 + sleep 10 + wait_job_completion required-some-name-required + local job=required-some-name-required + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.metadata.name}' required-some-name-required+ retry=0 ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/required-some-name-required -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete perconapgcluster some-name-required ++ mktemp + local LAST_OUT=/tmp/tmp.2Br41eLSCN ++ mktemp + local LAST_ERR=/tmp/tmp.P48hnxhTdN + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete perconapgcluster some-name-required + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.2Br41eLSCN perconapgcluster.pg.percona.com "some-name-required" deleted + cat /tmp/tmp.P48hnxhTdN + rm /tmp/tmp.2Br41eLSCN /tmp/tmp.P48hnxhTdN + return 0 + sleep 30 + cluster=some-name-disabled + spinup_pgcluster some-name-disabled /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml '' '' disabled + local cluster=some-name-disabled + local config=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local backup=false + local custom_config=false + local podAntiAffinity=disabled + local schedule=false + local schedule_type=false + local tls_only=false + local pgoClientFile=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + local namespace=affinity-12045 + local root_via_pgbouncer= + desc 'create fresh PG cluster' + set +o xtrace ----------------------------------------------------------------------------------- create fresh PG cluster ----------------------------------------------------------------------------------- + create_user_secrets some-name-disabled affinity-12045 + local cluster=some-name-disabled + local namespace=affinity-12045 + yq r -d0 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - stringData.some-name some-name_pass + yq w - metadata.name some-name-disabled-users-secret + kubectl_bin -n affinity-12045 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ajvqDAdu3R ++ mktemp + local LAST_ERR=/tmp/tmp.D974fCuTBO + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.ajvqDAdu3R secret/some-name-disabled-users-secret created + cat /tmp/tmp.D974fCuTBO + rm /tmp/tmp.ajvqDAdu3R /tmp/tmp.D974fCuTBO + return 0 + yq r -d1 /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/users-secret.yaml + yq w - metadata.name some-name-disabled-pmm-secret + kubectl_bin -n affinity-12045 apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.1D6FCwin0P ++ mktemp + local LAST_ERR=/tmp/tmp.GYzIFwgwYD + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.1D6FCwin0P secret/some-name-disabled-pmm-secret created + cat /tmp/tmp.GYzIFwgwYD + rm /tmp/tmp.1D6FCwin0P /tmp/tmp.GYzIFwgwYD + return 0 + kubectl_bin -n affinity-12045 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.Juqgltfcmh ++ mktemp + local LAST_ERR=/tmp/tmp.tkq7CjIo3a + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/client.yml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Juqgltfcmh deployment.apps/pg-client unchanged + cat /tmp/tmp.tkq7CjIo3a + rm /tmp/tmp.Juqgltfcmh /tmp/tmp.tkq7CjIo3a + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml some-name-disabled false false false false disabled false false false + local path=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml + local name=some-name-disabled + local backup=false + local restore_from=false + local restore_opts=false + local custom_config=false + local podAntiAffinity=disabled + local schedule=false + local schedule_type=false + local tls_only=false + local root_via_pgbouncer= + yq w - metadata.labels.crunchy-pgha-scope some-name-disabled + yq w - metadata.labels.deployment-name some-name-disabled + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/deploy/cr.yaml metadata.annotations.current-primary some-name-disabled + yq w - metadata.labels.pg-cluster some-name-disabled + yq w - metadata.name some-name-disabled + yq w - spec.clustername some-name-disabled + yq w - metadata.labels.name some-name-disabled + yq w - spec.database some-name-disabled + yq w - spec.keepData false + yq w - spec.name some-name-disabled + yq w - spec.pgBadger.image perconalab/percona-postgresql-operator:main-ppg14-pgbadger + yq w - spec.pmm.image perconalab/pmm-client:2.41.2 + yq w - spec.pgPrimary.image perconalab/percona-postgresql-operator:main-ppg14-postgres-ha + yq w - spec.backup.image perconalab/percona-postgresql-operator:main-ppg14-pgbackrest + yq w - spec.pgBouncer.image perconalab/percona-postgresql-operator:main-ppg14-pgbouncer + yq w - spec.backup.backrestRepoImage perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo + yq w - spec.secretsName some-name-disabled-users-secret + yq w - spec.user some-name + yq w - spec.keepBackups false + case ${podAntiAffinity} in + yq w - spec.backup.affinity.antiAffinityType disabled + yq w - spec.pgBouncer.affinity.antiAffinityType disabled + yq w /tmp/tmp.Hi0Eb7SlVS/cr.yaml spec.pgPrimary.affinity.antiAffinityType disabled + mv /tmp/tmp.Hi0Eb7SlVS/cr.podAffinity.yaml /tmp/tmp.Hi0Eb7SlVS/cr.yaml + case ${backup} in + [[ false != \f\a\l\s\e ]] + [[ false != \f\a\l\s\e ]] + [[ false == \t\r\u\e ]] + [[ '' == \t\r\u\e ]] + kubectl_bin apply -f /tmp/tmp.Hi0Eb7SlVS/cr.yaml -n affinity-12045 ++ mktemp + local LAST_OUT=/tmp/tmp.67x7EzfRIn ++ mktemp + local LAST_ERR=/tmp/tmp.X4OvG125fJ + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.Hi0Eb7SlVS/cr.yaml -n affinity-12045 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.67x7EzfRIn perconapgcluster.pg.percona.com/some-name-disabled created + cat /tmp/tmp.X4OvG125fJ + rm /tmp/tmp.67x7EzfRIn /tmp/tmp.X4OvG125fJ + return 0 + wait_deployment some-name-disabled-backrest-shared-repo + local name=some-name-disabled-backrest-shared-repo + local target_namespace=affinity-12045 + sleep 10 + set +o xtrace some-name-disabled-backrest-shared-repo... + wait_deployment some-name-disabled + local name=some-name-disabled + local target_namespace=affinity-12045 + sleep 10 + set +o xtrace some-name-disabled......... + [[ false == \g\c\s\+* ]] + wait_job_completion some-name-disabled-stanza-create false + local job=some-name-disabled-stanza-create + local mandatory=false + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.metadata.name}' some-name-disabled-stanza-create+ retry=0 ++ kubectl_bin -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/some-name-disabled-stanza-create -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ not found == \n\o\t\ \f\o\u\n\d ]] + echo + wait_job_completion backrest-backup-some-name-disabled + local job=backrest-backup-some-name-disabled + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' backrest-backup-some-name-disabled+ retry=0 ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/backrest-backup-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + wait_cluster_consistency some-name-disabled + cluster_name=some-name-disabled + wait_cluster_status some-name-disabled 'pgcluster Initialized' + local cluster_name=some-name-disabled + local 'target_status=pgcluster Initialized' + local target_namespace=affinity-12045 + local status_query= + case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in ++ kubectl_bin get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pcqwTekBGH +++ mktemp ++ local LAST_ERR=/tmp/tmp.6uToHxPzeK ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.pcqwTekBGH ++ cat /tmp/tmp.6uToHxPzeK ++ rm /tmp/tmp.pcqwTekBGH /tmp/tmp.6uToHxPzeK ++ return 0 + status_query='kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o jsonpath='\''{.status.PGCluster.state}'\''' + local retry=0 ++ eval kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.YgsbNDXOvy ++++ mktemp +++ local LAST_ERR=/tmp/tmp.KZUgH8cXCV +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.YgsbNDXOvy +++ cat /tmp/tmp.KZUgH8cXCV +++ rm /tmp/tmp.YgsbNDXOvy /tmp/tmp.KZUgH8cXCV +++ return 0 + [[ pgcluster Processed == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + let retry+=1 + '[' 1 -ge 24 ']' + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 10 ++ eval kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath='\''{.status.PGCluster.state}'\''' +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.status.PGCluster.state}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MMRrvoalix ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ppeVjjufbg +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.status.PGCluster.state}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MMRrvoalix +++ cat /tmp/tmp.ppeVjjufbg +++ rm /tmp/tmp.MMRrvoalix /tmp/tmp.ppeVjjufbg +++ return 0 + [[ pgcluster Initialized == \p\g\c\l\u\s\t\e\r\ \I\n\i\t\i\a\l\i\z\e\d ]] + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- ++ get_psql_user_pass postgres some-name-disabled ++ local user=postgres ++ local cluster=some-name-disabled ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.K6RzxiBYsi ++++ mktemp +++ local LAST_ERR=/tmp/tmp.97A2X86nJ6 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.K6RzxiBYsi +++ cat /tmp/tmp.97A2X86nJ6 +++ rm /tmp/tmp.K6RzxiBYsi /tmp/tmp.97A2X86nJ6 +++ return 0 ++ [[ xsome-name-disabled == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.XtIDCZTsZX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5TXxRV24kh +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.XtIDCZTsZX +++ cat /tmp/tmp.5TXxRV24kh +++ rm /tmp/tmp.XtIDCZTsZX /tmp/tmp.5TXxRV24kh +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WiZ0PAHVgQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.DnnnpL1pM8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.WiZ0PAHVgQ +++ cat /tmp/tmp.DnnnpL1pM8 +++ rm /tmp/tmp.WiZ0PAHVgQ /tmp/tmp.DnnnpL1pM8 +++ return 0 ++ [[ xsome-name-disabled-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.oyA5n4sGSr +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.NeuXM4WS8D ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.oyA5n4sGSr ++++ cat /tmp/tmp.NeuXM4WS8D ++++ rm /tmp/tmp.oyA5n4sGSr /tmp/tmp.NeuXM4WS8D ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-disabled-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0dOKxTiszm ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BGl2I6gjm5 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-disabled-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.0dOKxTiszm +++ cat /tmp/tmp.BGl2I6gjm5 +++ rm /tmp/tmp.0dOKxTiszm /tmp/tmp.BGl2I6gjm5 +++ return 0 ++ echo postgres_password + run_psql 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:postgres_password@some-name-disabled.affinity-12045 + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local uri=postgres:postgres_password@some-name-disabled.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.miP3MP1NMj +++ mktemp ++ local LAST_ERR=/tmp/tmp.D2uo0dhOtI ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.miP3MP1NMj ++ cat /tmp/tmp.D2uo0dhOtI ++ rm /tmp/tmp.miP3MP1NMj /tmp/tmp.D2uo0dhOtI ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-disabled.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.WL4Vo2Uohc ++ mktemp + local LAST_ERR=/tmp/tmp.C5KCgB9ToF + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-disabled.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.WL4Vo2Uohc + cat /tmp/tmp.C5KCgB9ToF + rm /tmp/tmp.WL4Vo2Uohc /tmp/tmp.C5KCgB9ToF + return 0 ++ get_psql_user_pass postgres some-name-disabled ++ local user=postgres ++ local cluster=some-name-disabled ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.FKgOy9L7h7 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Hx2ysVEBGM +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.FKgOy9L7h7 +++ cat /tmp/tmp.Hx2ysVEBGM +++ rm /tmp/tmp.FKgOy9L7h7 /tmp/tmp.Hx2ysVEBGM +++ return 0 ++ [[ xsome-name-disabled == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WdS3b4CbnX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.epcuLz0AW9 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.WdS3b4CbnX +++ cat /tmp/tmp.epcuLz0AW9 +++ rm /tmp/tmp.WdS3b4CbnX /tmp/tmp.epcuLz0AW9 +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.puuCy0mdZW ++++ mktemp +++ local LAST_ERR=/tmp/tmp.abSeyyMyaG +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.puuCy0mdZW +++ cat /tmp/tmp.abSeyyMyaG +++ rm /tmp/tmp.puuCy0mdZW /tmp/tmp.abSeyyMyaG +++ return 0 ++ [[ xsome-name-disabled-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.nWPyjnkr44 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.16aPCJ2Ngr ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.nWPyjnkr44 ++++ cat /tmp/tmp.16aPCJ2Ngr ++++ rm /tmp/tmp.nWPyjnkr44 /tmp/tmp.16aPCJ2Ngr ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-disabled-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.N8H1MBpWnX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.4yU9fUm4iH +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-disabled-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.N8H1MBpWnX +++ cat /tmp/tmp.4yU9fUm4iH +++ rm /tmp/tmp.N8H1MBpWnX /tmp/tmp.4yU9fUm4iH +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:postgres_password@some-name-disabled.affinity-12045 + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' + local uri=postgres:postgres_password@some-name-disabled.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HFgHEFaDoo +++ mktemp ++ local LAST_ERR=/tmp/tmp.nMV6hkwCZO ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.HFgHEFaDoo ++ cat /tmp/tmp.nMV6hkwCZO ++ rm /tmp/tmp.HFgHEFaDoo /tmp/tmp.nMV6hkwCZO ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-disabled.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.xP44okOZIO ++ mktemp + local LAST_ERR=/tmp/tmp.OAItm0j0uF + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-disabled.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.xP44okOZIO + cat /tmp/tmp.OAItm0j0uF + rm /tmp/tmp.xP44okOZIO /tmp/tmp.OAItm0j0uF + return 0 ++ get_psql_user_pass postgres some-name-disabled ++ local user=postgres ++ local cluster=some-name-disabled ++ local namespace=affinity-12045 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MtRxD7VYFi ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XGxPDy4nz8 +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.name}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MtRxD7VYFi +++ cat /tmp/tmp.XGxPDy4nz8 +++ rm /tmp/tmp.MtRxD7VYFi /tmp/tmp.XGxPDy4nz8 +++ return 0 ++ [[ xsome-name-disabled == \x ]] ++ case $(kubectl_bin get perconapgcluster/${cluster_name} -o jsonpath='{.metadata.labels.pgo-version}') in +++ kubectl_bin get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MQcUbXJvcv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gs3q5HwfKm +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl get perconapgcluster/some-name-disabled -o 'jsonpath={.metadata.labels.pgo-version}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.MQcUbXJvcv +++ cat /tmp/tmp.gs3q5HwfKm +++ rm /tmp/tmp.MQcUbXJvcv /tmp/tmp.gs3q5HwfKm +++ return 0 +++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.EUqwQiF6Lx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.areVbKhNDC +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.EUqwQiF6Lx +++ cat /tmp/tmp.areVbKhNDC +++ rm /tmp/tmp.EUqwQiF6Lx /tmp/tmp.areVbKhNDC +++ return 0 ++ [[ xsome-name-disabled-users-secret == \x ]] +++ base64 -d ++++ kubectl_bin -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.evJJdQcgR9 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.bKzI0YTNTs ++++ local exit_status=0 ++++ local timeout=4 ++++ set +o errexit +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++++ kubectl -n affinity-12045 get perconapgcluster/some-name-disabled -o 'jsonpath={.spec.secretsName}' ++++ exit_status=0 ++++ [[ 0 != 0 ]] ++++ break ++++ set -o errexit ++++ cat /tmp/tmp.evJJdQcgR9 ++++ cat /tmp/tmp.bKzI0YTNTs ++++ rm /tmp/tmp.evJJdQcgR9 /tmp/tmp.bKzI0YTNTs ++++ return 0 +++ kubectl_bin -n affinity-12045 get secret/some-name-disabled-users-secret -o 'jsonpath={.data.postgres}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tJCx1hWONX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.kRHKXHr90g +++ local exit_status=0 +++ local timeout=4 +++ set +o errexit ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox +++ kubectl -n affinity-12045 get secret/some-name-disabled-users-secret -o 'jsonpath={.data.postgres}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ set -o errexit +++ cat /tmp/tmp.tJCx1hWONX +++ cat /tmp/tmp.kRHKXHr90g +++ rm /tmp/tmp.tJCx1hWONX /tmp/tmp.kRHKXHr90g +++ return 0 ++ echo postgres_password + run_psql '\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' postgres:postgres_password@some-name-disabled.affinity-12045 + local 'command=\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";' + local uri=postgres:postgres_password@some-name-disabled.affinity-12045 + local driver=postgres + local suffix=.svc.cluster.local ++ kubectl_bin -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SYtlKUQfEc +++ mktemp ++ local LAST_ERR=/tmp/tmp.Tn4I5WwxrD ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl -n affinity-12045 get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.SYtlKUQfEc ++ cat /tmp/tmp.Tn4I5WwxrD ++ rm /tmp/tmp.SYtlKUQfEc /tmp/tmp.Tn4I5WwxrD ++ return 0 + local client_container=pg-client-76df9df876-j696t + kubectl_bin exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-disabled.affinity-12045.svc.cluster.local' ++ mktemp + local LAST_OUT=/tmp/tmp.Js3BCc4Prp ++ mktemp + local LAST_ERR=/tmp/tmp.fJd9iIudrR + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl exec pg-client-76df9df876-j696t -n affinity-12045 -- bash -c 'printf '\''\c myapp \\\ GRANT SELECT,INSERT ON myApp to "some-name";GRANT USAGE ON SCHEMA public TO "some-name";\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://postgres:postgres_password@some-name-disabled.affinity-12045.svc.cluster.local' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Js3BCc4Prp + cat /tmp/tmp.fJd9iIudrR + rm /tmp/tmp.Js3BCc4Prp /tmp/tmp.fJd9iIudrR + return 0 + sleep 10 + compare_kubectl deployment/some-name-disabled + local resource=deployment/some-name-disabled + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-disabled.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/deployment_some-name-disabled.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-disabled + yq d - metadata.managedFields + yq d - '**.creationTimestamp' + yq d - '**.namespace' ++ mktemp + yq d - metadata.resourceVersion + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + local LAST_OUT=/tmp/tmp.KyvNg6Wtpc + yq d - metadata.selfLink + yq d - 'metadata.annotations."kubernetes.io/psp"' ++ mktemp + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - metadata.deletionTimestamp + yq d - '**.creationTimestamp' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - '**.image' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**.clusterIP' + yq d - '**.clusterIPs' + yq d - '**.dataSource' + yq d - '**.procMount' + yq d - '**.storageClassName' + yq d - '**.finalizers' + yq d - spec.manualSelector + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - status + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.(name==suffix)' + yq d - '**."percona.com/*"' + yq d - '**.preemptionPolicy' + yq d - '**.(volumeMode==Filesystem).volumeMode' + local LAST_ERR=/tmp/tmp.bbXZlaPmL0 + local exit_status=0 + local timeout=4 + set +o errexit + yq d - '**.nodePort' + yq d - '**.enableServiceLinks' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.(name==NAMESPACE)' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.imagePullSecrets' + yq d - '**.controller-uid' + yq d - spec.nodeName + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**.ipFamilies' + yq d - '**.ipFamilyPolicy' + yq d - spec.volumeMode + yq d - '**.healthCheckNodePort' + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.(name==PODNAME)' + yq d - spec.podReplacementPolicy + yq d - '**.internalTrafficPolicy' + yq d - '**.volumeName' ++ seq 0 2 + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**.uid' + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-disabled + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.KyvNg6Wtpc + cat /tmp/tmp.bbXZlaPmL0 + rm /tmp/tmp.KyvNg6Wtpc /tmp/tmp.bbXZlaPmL0 + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-disabled.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-disabled.yml /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-disabled.yml + compare_kubectl deployment/some-name-disabled-backrest-shared-repo + local resource=deployment/some-name-disabled-backrest-shared-repo + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-disabled-backrest-shared-repo.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/deployment_some-name-disabled-backrest-shared-repo.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + kubectl_bin get -o yaml deployment/some-name-disabled-backrest-shared-repo + yq d - '**.creationTimestamp' + yq d - '**.namespace' + yq d - '**.uid' + yq d - metadata.resourceVersion + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - '**.creationTimestamp' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - metadata.selfLink + yq d - '**.image' + yq d - '**.clusterIP' + yq d - '**.storageClassName' + yq d - '**.clusterIPs' + yq d - '**."percona.com/*"' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**.dataSource' + yq d - '**.healthCheckNodePort' ++ mktemp + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + local LAST_OUT=/tmp/tmp.9RlvST1RxN + yq d - metadata.deletionTimestamp + yq d - '**.nodePort' + yq d - '**.finalizers' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**.volumeName' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**.procMount' + yq d - spec.nodeName + yq d - '**."volume.kubernetes.io/storage-provisioner"' ++ mktemp + yq d - spec.volumeMode + yq d - '**.preemptionPolicy' + yq d - '**.controller-uid' + local LAST_ERR=/tmp/tmp.XqiKhpC0PR + local exit_status=0 + local timeout=4 + set +o errexit + yq d - '**.(name==PODNAME)' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - spec.podReplacementPolicy + yq d - spec.manualSelector + yq d - metadata.managedFields + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**.internalTrafficPolicy' + yq d - '**.imagePullSecrets' + yq d - '**.enableServiceLinks' ++ seq 0 2 + yq d - status + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml deployment/some-name-disabled-backrest-shared-repo + yq d - '**.(name==suffix)' + yq d - '**.(name==NAMESPACE)' + yq d - '**.(name==COMMAND_OPTS)' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.ipFamilies' + yq d - '**.ipFamilyPolicy' + yq d - '**.allocateLoadBalancerNodePorts' + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.9RlvST1RxN + cat /tmp/tmp.XqiKhpC0PR + rm /tmp/tmp.9RlvST1RxN /tmp/tmp.XqiKhpC0PR + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-disabled-backrest-shared-repo.yml kind + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/deployment_some-name-disabled-backrest-shared-repo.yml /tmp/tmp.Hi0Eb7SlVS/deployment_some-name-disabled-backrest-shared-repo.yml + compare_kubectl job/backrest-backup-some-name-disabled + local resource=job/backrest-backup-some-name-disabled + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/job_backrest-backup-some-name-disabled.yml + local new_result=/tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml + [[ GKE == \O\P\E\N\S\H\I\F\T ]] + yq d - metadata.managedFields + yq d - '**.creationTimestamp' + yq d - '**.namespace' + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - '**.uid' + yq d - metadata.deletionTimestamp + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - 'spec.template.metadata.labels."batch.kubernetes.io/job-name"' + yq d - '**.creationTimestamp' + yq d - 'metadata.annotations."batch.kubernetes.io/job-tracking"' + yq d - '**."batch.kubernetes.io/controller-uid"' + yq d - '**.procMount' + yq d - '**.image' + yq d - '**.clusterIP' + yq d - '**.clusterIPs' + yq d - '**.dataSource' + yq d - spec.volumeMode + yq d - '**.storageClassName' + yq d - metadata.selfLink + yq d - spec.nodeName + yq d - metadata.resourceVersion + yq d - '**.healthCheckNodePort' + yq d - '**.ipFamilyPolicy' + yq d - '**.internalTrafficPolicy' + yq d - '**.allocateLoadBalancerNodePorts' + yq d - '**.ipFamilies' + yq d - '**.finalizers' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**.volumeName' + yq d - '**.imagePullSecrets' + yq d - '**.enableServiceLinks' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - status + yq d - spec.podReplacementPolicy + yq d - '**.(name==suffix)' + kubectl_bin get -o yaml job/backrest-backup-some-name-disabled + yq d - spec.manualSelector + yq d - '**.controller-uid' + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**.preemptionPolicy' + yq d - '**."percona.com/*"' + yq d - '**.(volumeMode==Filesystem).volumeMode' ++ mktemp + yq d - '**.nodePort' + local LAST_OUT=/tmp/tmp.qGfqJenInJ + yq d - '**.(name==NAMESPACE)' + yq d - '**.(name==COMMAND_OPTS)' + yq d - '**.(name==PODNAME)' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - 'metadata.ownerReferences.*.apiVersion' ++ mktemp + local LAST_ERR=/tmp/tmp.Fpx6bxvsDc + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get -o yaml job/backrest-backup-some-name-disabled + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.qGfqJenInJ + cat /tmp/tmp.Fpx6bxvsDc + rm /tmp/tmp.qGfqJenInJ /tmp/tmp.Fpx6bxvsDc + return 0 + case $(yq r ${new_result} 'kind') in ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml kind ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml metadata.annotations + [[ x{} == \x ]] ++ yq r /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml metadata.annotations + [[ x{} == \x\{\} ]] + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml metadata.annotations + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml metadata.generation + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml spec.completionMode + yq d -i /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml spec.suspend + diff -u /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/affinity/compare/job_backrest-backup-some-name-disabled.yml /tmp/tmp.Hi0Eb7SlVS/job_backrest-backup-some-name-disabled.yml + create_backup some-name-disabled disabled + local cluster=some-name-disabled + local bckp_prefix=disabled + local bckp_type=full + yq w /mnt/jenkins/workspace/cloud-pg-operator_PR-776/e2e-tests/conf/backup.yml metadata.labels.pg-cluster some-name-disabled + yq w - metadata.name disabled-some-name-disabled + yq w - spec.name disabled-some-name-disabled + yq w --style=single -- - spec.parameters.backrest-opts --type=full + kubectl_bin apply -f - + yq w - spec.parameters.pg-cluster some-name-disabled + yq w - spec.namespace affinity-12045 ++ kubectl_bin get pods --selector=name=some-name-disabled-backrest-shared-repo,pg-cluster=some-name-disabled -o 'jsonpath={.items[].metadata.name}' + yq w - spec.parameters.job-name disabled-some-name-disabled +++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.dV7AiziaJG ++ local LAST_OUT=/tmp/tmp.iWH8Cwszua ++ mktemp +++ mktemp + local LAST_ERR=/tmp/tmp.2WDzdGEbCt + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f - ++ local LAST_ERR=/tmp/tmp.68UrZFR5Yz ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=some-name-disabled-backrest-shared-repo,pg-cluster=some-name-disabled -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.iWH8Cwszua ++ cat /tmp/tmp.68UrZFR5Yz ++ rm /tmp/tmp.iWH8Cwszua /tmp/tmp.68UrZFR5Yz ++ return 0 + yq w - spec.parameters.podname some-name-disabled-backrest-shared-repo-6d97b67686-dqx82 + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.dV7AiziaJG pgtask.pg.percona.com/disabled-some-name-disabled created + cat /tmp/tmp.2WDzdGEbCt + rm /tmp/tmp.dV7AiziaJG /tmp/tmp.2WDzdGEbCt + return 0 + sleep 10 + wait_job_completion disabled-some-name-disabled + local job=disabled-some-name-disabled + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' disabled-some-name-disabled+ retry=0 ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ grep -io 'not found' ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/disabled-some-name-disabled -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + destroy affinity-12045 + grep -v level=debug + grep -v level=info ++ get_operator_pod ++ kubectl_bin get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' +++ mktemp + grep -v 'get backup status: Job.batch' + grep -v 'Getting tasks for pod' + tee /tmp/tmp.Hi0Eb7SlVS/operator.log + sort -u + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + grep -v 'the object has been modified' + grep -v 'Getting pods from source' ++ local LAST_OUT=/tmp/tmp.qtCr3tenTH +++ mktemp ++ local LAST_ERR=/tmp/tmp.F6qgdWh5jB ++ local exit_status=0 ++ local timeout=4 ++ set +o errexit +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox ++ kubectl get pods --selector=name=postgres-operator -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ set -o errexit ++ cat /tmp/tmp.qtCr3tenTH ++ cat /tmp/tmp.F6qgdWh5jB ++ rm /tmp/tmp.qtCr3tenTH /tmp/tmp.F6qgdWh5jB ++ return 0 + kubectl_bin logs postgres-operator-775b589b68-gf5sx -c operator ++ mktemp + local LAST_OUT=/tmp/tmp.3PZHRDKgsS ++ mktemp + local LAST_ERR=/tmp/tmp.3fnM74RVnK + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl logs postgres-operator-775b589b68-gf5sx -c operator + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.3PZHRDKgsS + cat /tmp/tmp.3fnM74RVnK + rm /tmp/tmp.3PZHRDKgsS /tmp/tmp.3fnM74RVnK + return 0 ] { } }, ] } }, { } }, ] { } }, { } ] } { } }, ] } ] { } }, }, { ] { } }, ] ] ], { } }, }, { }, { },{ }] }], ] { } }, , ], { } }, }], { } }, }], } ], } }, } }, } }, { } }, }, 2024/05/24 20:21:22 INF 1 (localhost:4150) connecting to nsqd 2024/05/24 20:22:10 INF 2 (localhost:4150) connecting to nsqd 2024/05/24 20:22:10 INF 3 (localhost:4150) connecting to nsqd 2024/05/24 20:22:17 INF 4 (localhost:4150) connecting to nsqd 2024/05/24 20:22:45 INF 5 (localhost:4150) connecting to nsqd 2024/05/24 20:22:45 INF 6 (localhost:4150) connecting to nsqd 2024/05/24 20:22:46 INF 7 (localhost:4150) connecting to nsqd 2024/05/24 20:24:10 INF 8 (localhost:4150) connecting to nsqd 2024/05/24 20:24:38 INF 9 (localhost:4150) connecting to nsqd 2024/05/24 20:24:39 INF 10 (localhost:4150) connecting to nsqd 2024/05/24 20:25:03 INF 11 (localhost:4150) connecting to nsqd 2024/05/24 20:25:08 INF 12 (localhost:4150) connecting to nsqd 2024/05/24 20:25:16 INF 13 (localhost:4150) connecting to nsqd 2024/05/24 20:26:05 INF 14 (localhost:4150) connecting to nsqd 2024/05/24 20:26:06 INF 15 (localhost:4150) connecting to nsqd 2024/05/24 20:26:13 INF 16 (localhost:4150) connecting to nsqd 2024/05/24 20:26:39 INF 17 (localhost:4150) connecting to nsqd 2024/05/24 20:26:39 INF 18 (localhost:4150) connecting to nsqd 2024/05/24 20:26:39 INF 19 (localhost:4150) connecting to nsqd 2024/05/24 20:28:03 INF 20 (localhost:4150) connecting to nsqd 2024/05/24 20:28:22 INF 21 (localhost:4150) connecting to nsqd 2024/05/24 20:28:24 INF 22 (localhost:4150) connecting to nsqd 2024/05/24 20:28:45 INF 23 (localhost:4150) connecting to nsqd 2024/05/24 20:28:50 INF 24 (localhost:4150) connecting to nsqd 2024/05/24 20:28:59 INF 25 (localhost:4150) connecting to nsqd 2024/05/24 20:29:48 INF 26 (localhost:4150) connecting to nsqd 2024/05/24 20:29:48 INF 27 (localhost:4150) connecting to nsqd 2024/05/24 20:29:55 INF 28 (localhost:4150) connecting to nsqd 2024/05/24 20:30:23 INF 29 (localhost:4150) connecting to nsqd 2024/05/24 20:30:23 INF 30 (localhost:4150) connecting to nsqd 2024/05/24 20:30:24 INF 31 (localhost:4150) connecting to nsqd 2024/05/24 20:31:41 INF 32 (localhost:4150) connecting to nsqd 2024/05/24 20:32:01 INF 33 (localhost:4150) connecting to nsqd "accessModes": [ "affinity": "affinity": { "affinity": { }, "affinity": { "podAntiAffinity": { "allowPrivilegeEscalation": false, "annotations": {"keep-backups":"false","keep-data":"false"}, "apiVersion": "apps/v1", "apiVersion": "batch/v1", "apiVersion": "v1", "apiVersion": "v1", "apiVersion": "v1", "backrest-command": "backup", "backrest-command": "stanza-create", "claimName": "some-name-disabled-pgbr-repo" "claimName": "some-name-pgbr-repo" "claimName": "some-name-required-pgbr-repo" "command": [ "configMap": { "configMap": { "configMapKeyRef": { "containerName": "database", "containerPort": 2022, "containerPort": 5432, "containerPort": 8009, "containers": [ "containers": [{ "cpu": "1", "cpu": "200m", "cpu": "500m", "crunchydata" "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "crunchy-pgbouncer": "true", "defaultMode": 288 "defaultMode": 288, "defaultMode": 288, "defaultMode": 420, "deployment-name": "some-name" "deployment-name": "some-name-disabled" "deployment-name": "some-name-disabled-repl1" "deployment-name": "some-name-disabled-repl1","pgo-version": "1.7.0","name": "some-name-disabled-replica","pg-cluster": "some-name-disabled" "deployment-name": "some-name-disabled-repl2" "deployment-name": "some-name-disabled-repl2","pgo-version": "1.7.0","name": "some-name-disabled-replica","pg-cluster": "some-name-disabled" "deployment-name": "some-name-repl1" "deployment-name": "some-name-repl2" "deployment-name": "some-name-required" "deployment-name": "some-name-required-repl1" "deployment-name": "some-name-required-repl2" "divisor": "1m", "dnsPolicy": "ClusterFirst" "downwardAPI": { "emptyDir": { "emptyDir": { "emptyDir": { "emptyDir": { "env": [ "env": [{ "exec": { "fieldPath": "metadata.annotations" "fieldPath": "metadata.labels" "fieldPath": "metadata.labels['crunchy-pgha-scope']" "fieldPath": "metadata.namespace" "fieldRef": { "fieldRef": { "fsGroup": 2, I0524 20:20:46.854643 1 shared_informer.go:240] Waiting for caches to sync for namespace I0524 20:20:46.955269 1 shared_informer.go:247] Caches are synced for namespace I0524 20:20:47.041419 1 shared_informer.go:240] Waiting for caches to sync for affinity-12045 I0524 20:20:47.142410 1 shared_informer.go:247] Caches are synced for affinity-12045 "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbackrest-repo", "image": "perconalab/percona-postgresql-operator:main-ppg14-pgbouncer", "image": "perconalab/percona-postgresql-operator:main-ppg14-postgres-ha", "image": "perconalab/percona-postgresql-operator:PR-776-0087ef129-pgo-rmdata", "imagePullPolicy": "IfNotPresent" "imagePullPolicy": "IfNotPresent", "initialDelaySeconds": 15 "initialDelaySeconds": 30, "items": [ "items": [ "items": [ "key": "config", "key": "config", "key": "crunchy-pgbouncer", "key": "init" "key": "password" "key": "pg-cluster", "key": "pgo-backrest", "key": "pgo-backrest-job", "key": "pgo-pg-database", "key": "pg-pod-anti-affinity", "key": "tls.crt", "key": "tls.key", "key": "vendor", "kind": "Deployment", "kind": "Job", "kind": "PersistentVolumeClaim", "kind": "Service", "labels": { "labels": { "labels": { "labelSelector": { "livenessProbe": { "matchExpressions": [ "matchLabels": { "matchLabels": { "maxSurge": 1 "maxUnavailable": 1, "medium": "Memory", "medium": "Memory" "medium": "Memory", "medium": "Memory", "medium": "Memory", "memory": "128Mi" "memory": "256Mi" "memory": "48Mi" "metadata": { "metadata": { "metadata": { "mountPath": "/backrestrepo", "mountPath": "/dev/shm", "mountPath": "/etc/pgbackrest" "mountPath": "/etc/pgbackrest/conf.d", "mountPath": "/etc/podinfo", "mountPath": "/etc/ssh", "mountPath": "/etc/ssh", "mountPath": "/home/pgbackrest" "mountPath": "/pgconf", "mountPath": "/pgconf/", "mountPath": "/pgconf/pgreplicator", "mountPath": "/pgconf/pgsuper", "mountPath": "/pgconf/pguser", "mountPath": "/pgconf/tls", "mountPath": "/pgconf/tls/pgbouncer", "mountPath": "/pgconf/tls-replication", "mountPath": "/pgdata", "mountPath": "/sshd", "mountPath": "/sshd", "mountPath": "/tmp", "mountPath": "/tmp", "mountPath": "/tmp" "name": "backrest", "name": "backrest-backup-some-name", "name": "backrest-backup-some-name-disabled", "name": "backrest-backup-some-name-required", "name": "backrestrepo", "name": "backrestrepo", "name": "BACKREST_SKIP_CREATE_STANZA", "name": "cluster1-replication-ssl-keypair", "name": "cluster1-ssl-ca" "name": "cluster1-ssl-ca" "name": "cluster1-ssl-keypair" "name": "cluster1-ssl-keypair" "name": "COMMAND", "name": "COMMAND_OPTS", "name": "database", "name": "disabled-some-name-disabled", "name": "dshm" "name": "dshm", "name": "ENABLE_SSHD", "name": "IS_BACKUP", "name": "IS_REPLICA", "name": "LD_PRELOAD", "name": "LD_PRELOAD", "name": "MODE", "name": "NAMESPACE", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_GROUP", "name": "NSS_WRAPPER_PASSWD", "name": "NSS_WRAPPER_PASSWD", "name": "patroni" "name": "PATRONI_KUBERNETES_LABELS", "name": "PATRONI_KUBERNETES_NAMESPACE", "name": "PATRONI_KUBERNETES_SCOPE_LABEL", "name": "PATRONI_LOG_LEVEL", "name": "PATRONI_POSTGRESQL_DATA_DIR", "name": "PATRONI_SCOPE", "name": "pgbackrest" "name": "pgbackrest-conf", "name": "pgbackrest-conf", "name": "pgbackrest-config" "name": "pgbackrest-config", "name": "PGBACKREST_DB_HOST", "name": "PGBACKREST_DB_PATH", "name": "PGBACKREST_DB_PATH", "name": "pgbackrest-home", "name": "pgbackrest-home", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_LOG_PATH", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_PORT", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_PG1_SOCKET_PATH", "name": "PGBACKREST_REPO1_HOST", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_PATH", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_REPO1_TYPE", "name": "PGBACKREST_STANZA", "name": "PGBACKREST_STANZA", "name": "pgbouncer" "name": "pgbouncer", "name": "pgbouncer-conf", "name": "pgbouncer-conf", "name": "PG_CLUSTER", "name": "pgconf-volume" "name": "pgconf-volume", "name": "pgdata", "name": "pgdata", "name": "PGHA_DATABASE", "name": "PGHA_INIT", "name": "PGHA_PASSWORD_TYPE", "name": "PGHA_PGBACKREST", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_GCS_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_LOCAL_S3_STORAGE", "name": "PGHA_PGBACKREST_S3_VERIFY_TLS", "name": "PGHA_PG_PORT", "name": "PGHA_REPLICA_REINIT_ON_START_FAIL", "name": "PGHA_SCOPE", "name": "PGHA_STANDBY", "name": "PGHA_SYNC_REPLICATION", "name": "PGHA_TLS_ENABLED", "name": "PGHA_TLS_ONLY", "name": "PGHA_USER", "name": "PGHOST", "name": "PG_PASSWORD", "name": "PG_PRIMARY_SERVICE_NAME", "name": "PITR_TARGET", "name": "podinfo" "name": "podinfo", "name": "PODNAME", "name": "postgres" "name": "preferred-some-name", "name": "primary-volume" "name": "primary-volume", "name": "REMOVE_BACKUP", "name": "REMOVE_DATA", "name": "REPLICA_NAME", "name": "report", "name": "required-some-name-required", "name": "rmdata", "name": "root-volume" "name": "root-volume", "name": "some-name", "name": "some-name", "name": "some-name-backrest-shared-repo", "name": "some-name-backrest-shared-repo" "name": "some-name-backrest-shared-repo", "name": "some-name-backrest-shared-repo", "name": "some-name-backrest-shared-repo", "name": "some-name-disabled", "name": "some-name-disabled", "name": "some-name-disabled-backrest-shared-repo", "name": "some-name-disabled-backrest-shared-repo" "name": "some-name-disabled-backrest-shared-repo", "name": "some-name-disabled-backrest-shared-repo", "name": "some-name-disabled-backrest-shared-repo", "name": "some-name-disabled-pgbouncer", "name": "some-name-disabled-pgbouncer", "name": "some-name-disabled-pgbouncer", "name": "some-name-disabled-pgbouncer", "name": "some-name-disabled-pgbouncer-cm" "name": "some-name-disabled-pgbouncer-secret", "name": "some-name-disabled-pgbouncer-secret", "name": "some-name-disabled-pgbr-repo", "name": "some-name-disabled-pgha-config", "name": "some-name-disabled-pgha-config", "name": "some-name-disabled-repl1", "name": "some-name-disabled-repl1", "name": "some-name-disabled-repl2", "name": "some-name-disabled-repl2", "name": "some-name-disabled-stanza-create", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer", "name": "some-name-pgbouncer-cm" "name": "some-name-pgbouncer-secret", "name": "some-name-pgbouncer-secret", "name": "some-name-pgbr-repo", "name": "some-name","pg-cluster": "some-name","deployment-name": "some-name","pgouser": "admin","crunchy-pgha-scope": "some-name","pgo-version": "1.7.0" "name": "some-name-pgha-config", "name": "some-name-pgha-config", "name": "some-name-repl1", "name": "some-name-repl1", "name": "some-name-repl2", "name": "some-name-repl2", "name": "some-name-replica","pg-cluster": "some-name","deployment-name": "some-name-repl2","pgo-version": "1.7.0" "name": "some-name-required", "name": "some-name-required", "name": "some-name-required-backrest-shared-repo", "name": "some-name-required-backrest-shared-repo" "name": "some-name-required-backrest-shared-repo", "name": "some-name-required-backrest-shared-repo", "name": "some-name-required-backrest-shared-repo", "name": "some-name-required-pgbouncer", "name": "some-name-required-pgbouncer", "name": "some-name-required-pgbouncer", "name": "some-name-required-pgbouncer", "name": "some-name-required-pgbouncer-cm" "name": "some-name-required-pgbouncer-secret", "name": "some-name-required-pgbouncer-secret", "name": "some-name-required-pgbr-repo", "name": "some-name-required","pg-cluster": "some-name-required","deployment-name": "some-name-required","pgouser": "admin","crunchy-pgha-scope": "some-name-required","pgo-version": "1.7.0" "name": "some-name-required-pgha-config", "name": "some-name-required-pgha-config", "name": "some-name-required-repl1", "name": "some-name-required-repl1", "name": "some-name-required-repl2", "name": "some-name-required-repl2", "name": "some-name-required-rmdata-gakw", "name": "some-name-required-rmdata-gakw", "name": "some-name-required-stanza-create", "name": "some-name-rmdata-iapu", "name": "some-name-rmdata-iapu", "name": "some-name-stanza-create", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "ssh-config", "name": "sshd", "name": "sshd", "name": "sshd", "name": "sshd", "name": "tls-pgbouncer" "name": "tls-pgbouncer", "name": "tls-replication" "name": "tls-replication", "name": "tls-server" "name": "tls-server", "name": "tmp" "name": "tmp" "name": "tmp", "name": "tmp", "name": "tmp", "name": "user-volume" "name": "user-volume", "nodePort": 0 "operator": "Exists" "operator": "In", "/opt/crunchy/bin/postgres-ha/health/pgha-liveness.sh" "/opt/crunchy/bin/postgres-ha/health/pgha-readiness.sh" "optional": true "path": "annotations" "path": "cpu_limit", "path": "cpu_request", "path": "labels" "path": "mem_limit", "path": "mem_request", "path": "ssh_config" "path": "ssh_config" "path": "tls-replication.crt" "path": "tls-replication.key" "periodSeconds": 15, "persistentVolumeClaim": { "persistentVolumeClaim":{"claimName":"some-name"} "persistentVolumeClaim":{"claimName":"some-name-disabled"} "persistentVolumeClaim":{"claimName":"some-name-disabled-repl1"} "persistentVolumeClaim":{"claimName":"some-name-disabled-repl2"} "persistentVolumeClaim":{"claimName":"some-name-repl1"} "persistentVolumeClaim":{"claimName":"some-name-repl2"} "persistentVolumeClaim":{"claimName":"some-name-required"} "persistentVolumeClaim":{"claimName":"some-name-required-repl1"} "persistentVolumeClaim":{"claimName":"some-name-required-repl2"} "pg-cluster": "some-name" "pg-cluster": "some-name", "pg-cluster": "some-name" "pg-cluster": "some-name", "pg-cluster": "some-name" "pg-cluster": "some-name", "pg-cluster": "some-name","deployment-name": "some-name-repl1","pgo-version": "1.7.0","name": "some-name-replica" "pg-cluster": "some-name-disabled" "pg-cluster": "some-name-disabled", "pg-cluster": "some-name-disabled" "pg-cluster": "some-name-disabled", "pg-cluster": "some-name-disabled" "pg-cluster": "some-name-disabled", "pg-cluster": "some-name-required" "pg-cluster": "some-name-required", "pg-cluster": "some-name-required" "pg-cluster": "some-name-required", "pg-cluster": "some-name-required" "pg-cluster": "some-name-required", "pg-cluster": "some-name-required","deployment-name": "some-name-required-repl2","pgo-version": "1.7.0","name": "some-name-required-replica" "pgo-backrest-job": "true", "pgo-backrest-repo": "true" "pgo-backrest-repo": "true", "pgo-backrest-repo": "true" "pgo-backrest": "true", "pgo-pg-database": "true", "pgo-pg-database": "true", "pgo-version": "1.7.0","name": "some-name-disabled","pg-cluster": "some-name-disabled","deployment-name": "some-name-disabled","pgouser": "admin","crunchy-pgha-scope": "some-name-disabled" "pgo-version": "1.7.0","name": "some-name-disabled","pg-cluster": "some-name-disabled","deployment-name": "some-name-disabled","pgouser": "admin","crunchy-pgha-scope": "some-name-disabled" "pgo-version": "1.7.0","name": "some-name-disabled-replica","pg-cluster": "some-name-disabled","deployment-name": "some-name-disabled-repl1" "pgo-version": "1.7.0","name": "some-name-disabled-replica","pg-cluster": "some-name-disabled","deployment-name": "some-name-disabled-repl2" "pgo-version": "1.7.0","name": "some-name","pg-cluster": "some-name","deployment-name": "some-name","pgouser": "admin","crunchy-pgha-scope": "some-name" "pgo-version": "1.7.0","name": "some-name-replica","pg-cluster": "some-name","deployment-name": "some-name-repl1" "pgo-version": "1.7.0","name": "some-name-replica","pg-cluster": "some-name","deployment-name": "some-name-repl2" "pgo-version": "1.7.0","name": "some-name-required","pg-cluster": "some-name-required","deployment-name": "some-name-required","pgouser": "admin","crunchy-pgha-scope": "some-name-required" "pgo-version": "1.7.0","name": "some-name-required-replica","pg-cluster": "some-name-required","deployment-name": "some-name-required-repl1" "pgo-version": "1.7.0","name": "some-name-required-replica","pg-cluster": "some-name-required","deployment-name": "some-name-required-repl1" "pgo-version": "1.7.0","name": "some-name-required-replica","pg-cluster": "some-name-required","deployment-name": "some-name-required-repl2" "pg-pod-anti-affinity": "disabled", "pg-pod-anti-affinity": "disabled", "pg-pod-anti-affinity": "preferred", "pg-pod-anti-affinity": "preferred", "pg-pod-anti-affinity": "required", "pg-pod-anti-affinity": "required", "pgrmdata": "true", "pgrmdata": "true", "podAffinityTerm": { "podAntiAffinity": { "podAntiAffinity": { "podAntiAffinity": { "port": 2022, "ports": [{ "ports": [{ "preferredDuringSchedulingIgnoredDuringExecution": [ "privileged": false, "projected": { "projected": { "projected": { "projected": { "sources": [] } "protocol": "TCP", "protocol": "TCP", "readinessProbe": { "readOnly": false "readOnly": false "readOnly": false "readOnlyRootFilesystem": true "readOnly": true "readOnly": true "readOnly": true "ReadWriteOnce" "replicas": 0, "replicas": 1, "replicas": 3, "requests": { "requests": { "require" "required", "requiredDuringSchedulingIgnoredDuringExecution": [ "resourceFieldRef": { "resource": "limits.cpu" "resource": "limits.memory" "resource": "requests.cpu" "resource": "requests.memory" "resources": { "resources": { "restartPolicy": "Always", "restartPolicy": "Never" "rollingUpdate": { "runAsNonRoot": true "secret": { "secret": { "secret": { "secret": { "secret": { "secretKeyRef": { "secretName": "some-name-backrest-repo-config" "secretName": "some-name-backrest-repo-config", "secretName": "some-name-backrest-repo-config" "secretName": "some-name-backrest-repo-config", "secretName": "some-name-disabled-backrest-repo-config" "secretName": "some-name-disabled-backrest-repo-config", "secretName": "some-name-disabled-backrest-repo-config" "secretName": "some-name-disabled-backrest-repo-config", "secretName": "some-name-disabled-postgres-secret" "secretName": "some-name-disabled-primaryuser-secret" "secretName": "some-name-disabled-some-name-secret" "secretName": "some-name-postgres-secret" "secretName": "some-name-primaryuser-secret" "secretName": "some-name-required-backrest-repo-config" "secretName": "some-name-required-backrest-repo-config", "secretName": "some-name-required-backrest-repo-config" "secretName": "some-name-required-backrest-repo-config", "secretName": "some-name-required-postgres-secret" "secretName": "some-name-required-primaryuser-secret" "secretName": "some-name-required-some-name-secret" "secretName": "some-name-some-name-secret" "securityContext": { "securityContext": { "securityContext": {}, "securityContext": {"fsGroup": 26,"supplementalGroups": [1001]}, "securityContext": {"runAsNonRoot":true,"fsGroup":26}, "selector": { "selector": { "serviceAccountName": "pgo-backrest", "serviceAccountName": "pgo-default", "serviceAccountName": "pgo-pg", "serviceAccountName": "pgo-target", "service-name": "some-name-backrest-shared-repo", "service-name": "some-name-backrest-shared-repo", "service-name": "some-name-disabled-backrest-shared-repo", "service-name": "some-name-disabled-backrest-shared-repo", "service-name": "some-name-disabled-pgbouncer", "service-name": "some-name-disabled-pgbouncer", "service-name": "some-name-disabled-pgbouncer", "service-name": "some-name-pgbouncer", "service-name": "some-name-pgbouncer", "service-name": "some-name-pgbouncer", "service-name": "some-name-required-backrest-shared-repo", "service-name": "some-name-required-backrest-shared-repo", "service-name": "some-name-required-pgbouncer", "service-name": "some-name-required-pgbouncer", "service-name": "some-name-required-pgbouncer", "sessionAffinity": "None" "sizeLimit": "128Ki" "sizeLimit": "16Mi" "sizeLimit": "1Mi" "sizeLimit": "2Mi" "sizeLimit": "64Mi" "sizeLimit": "64Mi" "some-name" "some-name-required" "sources": [ "sources": [ "sources": [ "spec": { "spec": { "spec": { "storage": "1Gi" "strategy": { "targetPort": 2022, "template": { time="2024-05-24T20:20:46Z" level=warning msg="PGAdminStorage setting not set, will use PrimaryStorage setting" func="github.com/percona/percona-postgresql-operator/internal/config.(*PgoConfig).Validate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/config/pgoconfig.go:337" version=1.7.0 time="2024-05-24T20:21:27Z" level=error msg="get deployment some-name-repl1: deployments.apps \"some-name-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T20:21:32Z" level=error msg="get deployment some-name-repl2: deployments.apps \"some-name-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T20:22:10Z" level=error msg="Error completing workflow some-name-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T20:22:10Z" level=error msg="pgtasks.pg.percona.com \"some-name-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T20:22:46Z" level=warning msg="crv1 pgreplica some-name-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T20:22:46Z" level=warning msg="crv1 pgreplica some-name-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T20:22:47Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T20:22:47Z" level=error msg="pod some-name-repl1-5489f574fb-6phbb does not have a host assigned" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T20:22:47Z" level=error msg="pod some-name-repl1-5489f574fb-6phbb does not have a host assigned" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T20:22:47Z" level=error msg="pod some-name-repl2-759b4f7fdd-b84j5 does not have a host assigned" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T20:22:47Z" level=error msg="pod some-name-repl2-759b4f7fdd-b84j5 does not have a host assigned" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T20:22:47Z" level=warning msg="crv1 pgreplica some-name-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T20:24:39Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:24:42Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:24:43Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:24:44Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:24:45Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:24:47Z" level=error msg="ConfigMap Controller: cannot find pgcluster for configMap some-name-pgha-config (namespace affinity-12045),ignoring" func="github.com/percona/percona-postgresql-operator/internal/controller/configmap.(*Controller).handleConfigMapSync()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/configmap/synchandler.go:55" version=1.7.0 time="2024-05-24T20:24:51Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:24:55Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:24:56Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:25:00Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:25:02Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:25:03Z" level=error msg="pgclusters.pg.percona.com \"some-name\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:25:08Z" level=error msg="jobs.batch \"some-name-rmdata-iapu\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T20:25:19Z" level=error msg="get deployment some-name-required-repl1: deployments.apps \"some-name-required-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T20:25:24Z" level=error msg="get deployment some-name-required-repl2: deployments.apps \"some-name-required-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T20:26:05Z" level=error msg="Error completing workflow some-name-required-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T20:26:05Z" level=error msg="pgtasks.pg.percona.com \"some-name-required-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T20:26:39Z" level=warning msg="crv1 pgreplica some-name-required-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T20:26:39Z" level=warning msg="crv1 pgreplica some-name-required-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T20:26:47Z" level=error func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).GetLocalConfigFromCluster()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:332" version=1.7.0 time="2024-05-24T20:26:47Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/kubeapi.ExecToPodThroughAPI()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/kubeapi/exec.go:76" version=1.7.0 time="2024-05-24T20:26:47Z" level=error msg="unable to upgrade connection: container not found (\"database\")" func="github.com/percona/percona-postgresql-operator/internal/operator/config.(*LocalDB).Sync.func2()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/config/localdb.go:173" version=1.7.0 time="2024-05-24T20:28:24Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:25Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:26Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:27Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:32Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:36Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:37Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:41Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:43Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:45Z" level=error msg="pgclusters.pg.percona.com \"some-name-required\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/pod.(*Controller).onUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/pod/podcontroller.go:103" version=1.7.0 time="2024-05-24T20:28:50Z" level=error msg="jobs.batch \"some-name-required-rmdata-gakw\" not found" func="github.com/percona/percona-postgresql-operator/internal/controller/job.(*Controller).handleRMDataUpdate()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/controller/job/rmdatahandler.go:61" version=1.7.0 time="2024-05-24T20:29:03Z" level=error msg="get deployment some-name-disabled-repl1: deployments.apps \"some-name-disabled-repl1\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T20:29:08Z" level=error msg="get deployment some-name-disabled-repl2: deployments.apps \"some-name-disabled-repl2\" not found" func="github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica.CreateReplicaResource()" file="/go/src/github.com/percona/percona-postgresql-operator/percona/controllers/pgreplica/pgreplica.go:40" version=1.7.0 time="2024-05-24T20:29:48Z" level=error msg="Error completing workflow some-name-disabled-createcluster" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:48" version=1.7.0 time="2024-05-24T20:29:48Z" level=error msg="pgtasks.pg.percona.com \"some-name-disabled-createcluster\" not found" func="github.com/percona/percona-postgresql-operator/internal/operator/task.completeWorkflow()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/task/workflow.go:49" version=1.7.0 time="2024-05-24T20:30:24Z" level=warning msg="crv1 pgreplica some-name-disabled-repl1 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 time="2024-05-24T20:30:24Z" level=warning msg="crv1 pgreplica some-name-disabled-repl2 is already marked complete, will not recreate" func="github.com/percona/percona-postgresql-operator/internal/operator/cluster.ScaleBase()" file="/go/src/github.com/percona/percona-postgresql-operator/internal/operator/cluster/cluster.go:535" version=1.7.0 "timeoutSeconds": 10 "topologyKey": "kubernetes.io/hostname" "type": "ClusterIP", "type": "RollingUpdate", "value": "" "value": "5432" "value": "5432" "value": "/backrestrepo/some-name-backrest-shared-repo" "value": "/backrestrepo/some-name-backrest-shared-repo" "value": "/backrestrepo/some-name-disabled-backrest-shared-repo" "value": "/backrestrepo/some-name-disabled-backrest-shared-repo" "value": "/backrestrepo/some-name-required-backrest-shared-repo" "value": "/backrestrepo/some-name-required-backrest-shared-repo" "value": "backup" "value": "crunchy-pgha-scope" "value": "db" "value": "db" "value": " --db-host=10.87.32.39 --db-path=/pgdata/some-name" "value": " --db-host=10.87.32.43 --db-path=/pgdata/some-name-required" "value": " --db-host=10.87.32.47 --db-path=/pgdata/some-name-disabled" "value": "false" "value": "false" "valueFrom": { "value": "INFO" "value": "pgbackrest" "value": "pgbackrest-repo" "value": "/pgdata/some-name" "value": "/pgdata/some-name" "value": "/pgdata/some-name-disabled" "value": "/pgdata/some-name-disabled" "value": "/pgdata/some-name-disabled-repl1" "value": "/pgdata/some-name-disabled-repl1" "value": "/pgdata/some-name-disabled-repl2" "value": "/pgdata/some-name-disabled-repl2" "value": "/pgdata/some-name-repl1" "value": "/pgdata/some-name-repl1" "value": "/pgdata/some-name-repl2" "value": "/pgdata/some-name-repl2" "value": "/pgdata/some-name-required" "value": "/pgdata/some-name-required" "value": "/pgdata/some-name-required-repl1" "value": "/pgdata/some-name-required-repl1" "value": "/pgdata/some-name-required-repl2" "value": "/pgdata/some-name-required-repl2" "value": "posix" "value": "posix" "value": "postgres" "values": [ "value": "some-name" "value": "some-name-backrest-shared-repo" "value": "some-name-backrest-shared-repo-6f8f947dbc-96d8f" "value": "some-name-disabled" "value": "some-name-disabled-backrest-shared-repo" "value": "some-name-disabled-backrest-shared-repo-6d97b67686-dqx82" "value": "some-name-required" "value": "some-name-required-backrest-shared-repo" "value": "some-name-required-backrest-shared-repo-8558c9b6fc-2cdk9" "value": "stanza-create" "value": "/tmp" "value": "/tmp" "value": "/tmp/nss_wrapper/pgbackrest-repo/group" "value": "/tmp/nss_wrapper/pgbackrest-repo/passwd" "value": "/tmp/nss_wrapper/postgres/group" "value": "/tmp/nss_wrapper/postgres/passwd" "value": "true" "value": "true" "value": "--type=full --db-host=10.87.32.39 --db-path=/pgdata/some-name" "value": "--type=full --db-host=10.87.32.43 --db-path=/pgdata/some-name-required" "value": "--type=full --db-host=10.87.32.47 --db-path=/pgdata/some-name-disabled" "value": "/usr/lib64/libnss_wrapper.so" "value": "/usr/lib64/libnss_wrapper.so" "value": "{vendor: \"crunchydata\"}" "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "vendor": "crunchydata" "vendor": "crunchydata", "volumeMounts": [ "volumeMounts": [{ "volumes": [ "volumes": [{ "weight": 1, + destroy_operator + kubectl_bin delete -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.LoXn1Bu2L1 ++ mktemp + local LAST_ERR=/tmp/tmp.yLN5vyqKse + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.Hi0Eb7SlVS/operator.yaml + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.LoXn1Bu2L1 + cat /tmp/tmp.yLN5vyqKse Error from server (NotFound): error when deleting "/tmp/tmp.Hi0Eb7SlVS/operator.yaml": serviceaccounts "pgo-deployer-sa" not found Error from server (NotFound): error when deleting "/tmp/tmp.Hi0Eb7SlVS/operator.yaml": clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found Error from server (NotFound): error when deleting "/tmp/tmp.Hi0Eb7SlVS/operator.yaml": configmaps "pgo-deployer-cm" not found Error from server (NotFound): error when deleting "/tmp/tmp.Hi0Eb7SlVS/operator.yaml": clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-crb" not found Error from server (NotFound): error when deleting "/tmp/tmp.Hi0Eb7SlVS/operator.yaml": jobs.batch "pgo-deploy" not found + rm /tmp/tmp.LoXn1Bu2L1 /tmp/tmp.yLN5vyqKse + return 1 + true + yq w -d4 /tmp/tmp.Hi0Eb7SlVS/operator.yaml 'spec.template.spec.containers[0].env[0].value' uninstall + kubectl_bin apply -f /tmp/tmp.Hi0Eb7SlVS/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.LZ2orh9AuL ++ mktemp + local LAST_ERR=/tmp/tmp.cNgXHbJzML + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl apply -f /tmp/tmp.Hi0Eb7SlVS/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.LZ2orh9AuL serviceaccount/pgo-deployer-sa created clusterrole.rbac.authorization.k8s.io/pgo-deployer-cr created configmap/pgo-deployer-cm created clusterrolebinding.rbac.authorization.k8s.io/pgo-deployer-crb created job.batch/pgo-deploy created + cat /tmp/tmp.cNgXHbJzML + rm /tmp/tmp.LZ2orh9AuL /tmp/tmp.cNgXHbJzML + return 0 + wait_job_completion pgo-deploy true + local job=pgo-deploy + local mandatory=true + local ns=affinity-12045 + retry=0 + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' pgo-deploy+ retry=0 ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 1 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 2 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 3 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 4 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 5 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 6 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 7 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 8 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 9 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 10 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 11 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 12 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 13 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 14 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 15 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 16 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 17 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 18 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 19 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 20 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 21 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 22 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 23 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 24 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 25 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 26 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 27 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 28 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 29 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 30 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 31 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 32 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 33 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 34 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 35 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 36 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 37 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 38 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 39 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ '' == \1 ]] + sleep 1 + echo -n . .+ let retry+=1 + '[' 40 -ge 600 ']' ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.metadata.name}' ++ grep -io 'not found' + [[ '' == \n\o\t\ \f\o\u\n\d ]] ++ kubectl_bin -n affinity-12045 get job/pgo-deploy -o 'jsonpath={.status.succeeded}' + [[ 1 == \1 ]] + echo + kubectl_bin delete -f /tmp/tmp.Hi0Eb7SlVS/operator-destroy.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.Od3xV3Gj7P ++ mktemp + local LAST_ERR=/tmp/tmp.9dYUJdPyro + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete -f /tmp/tmp.Hi0Eb7SlVS/operator-destroy.yaml + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.Od3xV3Gj7P serviceaccount "pgo-deployer-sa" deleted clusterrole.rbac.authorization.k8s.io "pgo-deployer-cr" deleted configmap "pgo-deployer-cm" deleted clusterrolebinding.rbac.authorization.k8s.io "pgo-deployer-crb" deleted job.batch "pgo-deploy" deleted + cat /tmp/tmp.9dYUJdPyro + rm /tmp/tmp.Od3xV3Gj7P /tmp/tmp.9dYUJdPyro + return 0 + egrep -v '^kube-|^default|Terminating|openshift|^NAME' + awk '{print$1}' + kubectl_bin get ns + xargs kubectl delete ns --grace-period=0 --force=true ++ mktemp + local LAST_OUT=/tmp/tmp.LiKqKILm9c ++ mktemp + local LAST_ERR=/tmp/tmp.UvKySzUtzT + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + set -o errexit + cat /tmp/tmp.LiKqKILm9c + cat /tmp/tmp.UvKySzUtzT + rm /tmp/tmp.LiKqKILm9c /tmp/tmp.UvKySzUtzT + return 0 Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. namespace "affinity-12045" force deleted + kubectl_bin delete clusterrolebindings pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.YJmRWWCvkS ++ mktemp + local LAST_ERR=/tmp/tmp.3EXHCQVsm3 + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterrolebindings pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.YJmRWWCvkS + cat /tmp/tmp.3EXHCQVsm3 Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.YJmRWWCvkS /tmp/tmp.3EXHCQVsm3 + return 1 + true + kubectl_bin delete clusterroles pgo-cluster-role pgo-deployer-cr ++ mktemp + local LAST_OUT=/tmp/tmp.xWXQD7iH4w ++ mktemp + local LAST_ERR=/tmp/tmp.8uItJaIoHu + local exit_status=0 + local timeout=4 + set +o errexit ++ seq 0 2 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 4 + for i in '$(seq 0 2)' + KUBECONFIG=/tmp/jen-pg-776-0087ef129-2-sandbox + kubectl delete clusterroles pgo-cluster-role pgo-deployer-cr + exit_status=1 + [[ 1 != 0 ]] + sleep 8 + set -o errexit + cat /tmp/tmp.xWXQD7iH4w + cat /tmp/tmp.8uItJaIoHu Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-cluster-role" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pgo-deployer-cr" not found + rm /tmp/tmp.xWXQD7iH4w /tmp/tmp.8uItJaIoHu + return 1 + true + rm -rf /tmp/tmp.Hi0Eb7SlVS