++ echo 'Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/logs/users.log' Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/logs/users.log ++ '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP= ++ oc get projects error: No Auth Provider found for name "gcp" +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep '\-eks\-' ++ '[' ']' ++ EKS=0 +++ kubectl version -o json +++ /usr/bin/sed -r 's/[^0-9.]+//g' +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' ++ KUBE_VERSION=1.20 +++ helm version -c +++ /usr/bin/sed -re 's/.*SemVer:"([^"]+)".*/\1/; s/.*\bVersion:"([^"]+)".*/\1/' ++ HELM_VERSION=v3.8.1 ++ '[' v3 == v2 ']' + create_infra users-19289 + local ns=users-19289 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name, label selector, or --all flag specified + : + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.SAZIN5QuvD ++ mktemp + local LAST_ERR=/tmp/tmp.se0Z5xxcui + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.SAZIN5QuvD No resources found + cat /tmp/tmp.se0Z5xxcui + rm /tmp/tmp.SAZIN5QuvD /tmp/tmp.se0Z5xxcui + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.GULTnRtHf4 ++ mktemp + local LAST_ERR=/tmp/tmp.xRGYVioybd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.GULTnRtHf4 No resources found + cat /tmp/tmp.xRGYVioybd + rm /tmp/tmp.GULTnRtHf4 /tmp/tmp.xRGYVioybd + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.PIW9o6xz3d ++ mktemp + local LAST_ERR=/tmp/tmp.EkgUN9O9cN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.PIW9o6xz3d No resources found + cat /tmp/tmp.EkgUN9O9cN + rm /tmp/tmp.PIW9o6xz3d /tmp/tmp.EkgUN9O9cN + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + kubectl_bin get ns + '[' '!' -z '' ']' + kubectl_bin delete namespace pxc-operator + xargs kubectl delete ns + awk '{print$1}' + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^NAME' ++ mktemp + local LAST_OUT=/tmp/tmp.luYGBNZqe6 ++ mktemp + local LAST_ERR=/tmp/tmp.893oBxtTYu + local exit_status=0 ++ seq 0 2 ++ mktemp + for i in '$(seq 0 2)' + kubectl get ns + local LAST_OUT=/tmp/tmp.D965RbCyXD ++ mktemp + local LAST_ERR=/tmp/tmp.vvEj9P6lOR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete namespace pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.luYGBNZqe6 + cat /tmp/tmp.893oBxtTYu + rm /tmp/tmp.luYGBNZqe6 /tmp/tmp.893oBxtTYu + return 0 error: resource(s) were provided, but no name, label selector, or --all flag specified + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.D965RbCyXD namespace "pxc-operator" deleted + cat /tmp/tmp.vvEj9P6lOR + rm /tmp/tmp.D965RbCyXD /tmp/tmp.vvEj9P6lOR + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + set +o xtrace namespace/pxc-operator - Error from server (NotFound): namespaces "pxc-operator" not found + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.XQ2SqL5uea ++ mktemp + local LAST_ERR=/tmp/tmp.wsE1nquqbR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl create namespace pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.XQ2SqL5uea namespace/pxc-operator created + cat /tmp/tmp.wsE1nquqbR + rm /tmp/tmp.XQ2SqL5uea /tmp/tmp.wsE1nquqbR + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.M5DVADbxml +++ mktemp ++ local LAST_ERR=/tmp/tmp.U2DBxOGkbk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.M5DVADbxml ++ cat /tmp/tmp.U2DBxOGkbk ++ rm /tmp/tmp.M5DVADbxml /tmp/tmp.U2DBxOGkbk ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.dqyQKqsbK6 ++ mktemp + local LAST_ERR=/tmp/tmp.XgofgtJhgi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.dqyQKqsbK6 Context "gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic" modified. + cat /tmp/tmp.XgofgtJhgi + rm /tmp/tmp.dqyQKqsbK6 /tmp/tmp.XgofgtJhgi + return 0 + deploy_operator + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.HtzyJZTyM4 ++ mktemp + local LAST_ERR=/tmp/tmp.prHX8NWfDL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/crd.yaml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.HtzyJZTyM4 customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbbackups.pxc.percona.com configured + cat /tmp/tmp.prHX8NWfDL + rm /tmp/tmp.HtzyJZTyM4 /tmp/tmp.prHX8NWfDL + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/cw-rbac.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.qwDxB1DuXZ ++ mktemp + local LAST_ERR=/tmp/tmp.vJWEPcqdvK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.qwDxB1DuXZ clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.vJWEPcqdvK + rm /tmp/tmp.qwDxB1DuXZ /tmp/tmp.vJWEPcqdvK + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/cw-operator.yaml + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.T6cfiA1UBc ++ mktemp + local LAST_ERR=/tmp/tmp.VOfUzg54ld + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.T6cfiA1UBc deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.VOfUzg54ld + rm /tmp/tmp.T6cfiA1UBc /tmp/tmp.VOfUzg54ld + return 0 + sleep 10 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.rZhe5IiosP +++ mktemp ++ local LAST_ERR=/tmp/tmp.D1tbQbFvIT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.rZhe5IiosP ++ cat /tmp/tmp.D1tbQbFvIT ++ rm /tmp/tmp.rZhe5IiosP /tmp/tmp.D1tbQbFvIT ++ return 0 + wait_pod percona-xtradb-cluster-operator-5699d7755d-kpc6q 480 pxc-operator + local pod=percona-xtradb-cluster-operator-5699d7755d-kpc6q + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-5699d7755d-kpc6q ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace percona-xtradb-cluster-operator-5699d7755d-kpc6q.Ok + sleep 3 + create_namespace users-19289 + local namespace=users-19289 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + '[' '!' -z '' ']' + kubectl_bin delete namespace users-19289 ++ mktemp + kubectl_bin get ns + local LAST_OUT=/tmp/tmp.eukzNKeaHC ++ mktemp + local LAST_OUT=/tmp/tmp.RsCq5vpjFt ++ mktemp + local LAST_ERR=/tmp/tmp.xgMrK94uPl + local exit_status=0 + awk '{print$1}' ++ mktemp + xargs kubectl delete ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^NAME' + local LAST_ERR=/tmp/tmp.bHBJFZqGqR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete namespace users-19289 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.RsCq5vpjFt + cat /tmp/tmp.bHBJFZqGqR + rm /tmp/tmp.RsCq5vpjFt /tmp/tmp.bHBJFZqGqR + return 0 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + kubectl delete namespace users-19289 error: resource(s) were provided, but no name, label selector, or --all flag specified + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + kubectl delete namespace users-19289 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + cat /tmp/tmp.eukzNKeaHC + cat /tmp/tmp.xgMrK94uPl Error from server (NotFound): namespaces "users-19289" not found + rm /tmp/tmp.eukzNKeaHC /tmp/tmp.xgMrK94uPl + return 1 + : + wait_for_delete namespace/users-19289 + local res=namespace/users-19289 + set +o xtrace namespace/users-19289 - Error from server (NotFound): namespaces "users-19289" not found + kubectl_bin create namespace users-19289 ++ mktemp + local LAST_OUT=/tmp/tmp.NoO0DgOjyO ++ mktemp + local LAST_ERR=/tmp/tmp.U2BoPyAkcq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl create namespace users-19289 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.NoO0DgOjyO namespace/users-19289 created + cat /tmp/tmp.U2BoPyAkcq + rm /tmp/tmp.NoO0DgOjyO /tmp/tmp.U2BoPyAkcq + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ub0jP0xX1T +++ mktemp ++ local LAST_ERR=/tmp/tmp.OxfQm8hULt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.Ub0jP0xX1T ++ cat /tmp/tmp.OxfQm8hULt ++ rm /tmp/tmp.Ub0jP0xX1T /tmp/tmp.OxfQm8hULt ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=users-19289 ++ mktemp + local LAST_OUT=/tmp/tmp.VNcjbKp5Jg ++ mktemp + local LAST_ERR=/tmp/tmp.ZLDxD2ooaL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=users-19289 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.VNcjbKp5Jg Context "gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic" modified. + cat /tmp/tmp.ZLDxD2ooaL + rm /tmp/tmp.VNcjbKp5Jg /tmp/tmp.ZLDxD2ooaL + return 0 + apply_secrets + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.5Xk4PNks4y ++ mktemp + local LAST_ERR=/tmp/tmp.zyNszRKBVe + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.5Xk4PNks4y secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.zyNszRKBVe + rm /tmp/tmp.5Xk4PNks4y /tmp/tmp.zyNszRKBVe + return 0 + desc 'create PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- + newpass=test-password ++ base64 ++ echo -n test-password + newpassencrypted=dGVzdC1wYXNzd29yZA== + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/client.yml + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.gs3RiWSFjA ++ mktemp + local LAST_ERR=/tmp/tmp.tgEa9NHija + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/secrets.yml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.gs3RiWSFjA secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.tgEa9NHija + rm /tmp/tmp.gs3RiWSFjA /tmp/tmp.tgEa9NHija + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/client.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.l4QFTgyBfJ + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1-11-0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-19289~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + /usr/bin/sed -e 's#initImage:.*-init$#initImage: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/client.yml + local LAST_ERR=/tmp/tmp.IbpA2jPRD4 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.l4QFTgyBfJ deployment.apps/pxc-client created + cat /tmp/tmp.IbpA2jPRD4 + rm /tmp/tmp.l4QFTgyBfJ /tmp/tmp.IbpA2jPRD4 + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/some-name.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.fnHTZQJryp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1-11-0#' ++ mktemp + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-19289~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_ERR=/tmp/tmp.fNmocsbAMM + local exit_status=0 + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#initImage:.*-init$#initImage: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.fnHTZQJryp perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.fNmocsbAMM + rm /tmp/tmp.fnHTZQJryp /tmp/tmp.fNmocsbAMM + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.gm4xa5bbHQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1IDQPQChKq +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ cat /tmp/tmp.gm4xa5bbHQ +++ cat /tmp/tmp.1IDQPQChKq +++ rm /tmp/tmp.gm4xa5bbHQ /tmp/tmp.1IDQPQChKq +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qY8O2Go6Gj ++++ mktemp +++ local LAST_ERR=/tmp/tmp.4k6WDsXnrU +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ cat /tmp/tmp.qY8O2Go6Gj +++ cat /tmp/tmp.4k6WDsXnrU +++ rm /tmp/tmp.qY8O2Go6Gj /tmp/tmp.4k6WDsXnrU +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace some-name-proxysql-0........Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace some-name-pxc-0........................Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace some-name-pxc-1.....................................Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace some-name-pxc-2..........................................Ok + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Nwu6AAooie +++ mktemp ++ local LAST_ERR=/tmp/tmp.EKUb9DHWfe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.Nwu6AAooie ++ cat /tmp/tmp.EKUb9DHWfe ++ rm /tmp/tmp.Nwu6AAooie /tmp/tmp.EKUb9DHWfe ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-5d749ff8b6-bdr9g + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bUAOZLdQM8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.mDb0sFj0y9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.bUAOZLdQM8 ++ cat /tmp/tmp.mDb0sFj0y9 ++ rm /tmp/tmp.bUAOZLdQM8 /tmp/tmp.mDb0sFj0y9 ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yTZq02OtHc +++ mktemp ++ local LAST_ERR=/tmp/tmp.vCJFotDbbW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.yTZq02OtHc ++ cat /tmp/tmp.vCJFotDbbW ++ rm /tmp/tmp.yTZq02OtHc /tmp/tmp.vCJFotDbbW ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1.sql /tmp/tmp.Ri7icm2AFE/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bKJJVitb1K +++ mktemp ++ local LAST_ERR=/tmp/tmp.O0nLfT7lA0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.bKJJVitb1K ++ cat /tmp/tmp.O0nLfT7lA0 ++ rm /tmp/tmp.bKJJVitb1K /tmp/tmp.O0nLfT7lA0 ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1.sql /tmp/tmp.Ri7icm2AFE/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zYqOERmGHt +++ mktemp ++ local LAST_ERR=/tmp/tmp.LUXrDxyg2N ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.zYqOERmGHt ++ cat /tmp/tmp.LUXrDxyg2N ++ rm /tmp/tmp.zYqOERmGHt /tmp/tmp.LUXrDxyg2N ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-1.sql /tmp/tmp.Ri7icm2AFE/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rWedVAnMW6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jquezFjf3b ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.rWedVAnMW6 ++ cat /tmp/tmp.jquezFjf3b Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.rWedVAnMW6 /tmp/tmp.jquezFjf3b ++ return 0 + '[' '' ']' + desc 'test root' + set +o xtrace ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets root dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=root + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.l4nFMDf7Bx ++ mktemp + local LAST_ERR=/tmp/tmp.sYBXZydhn8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.l4nFMDf7Bx secret/my-cluster-secrets patched + cat /tmp/tmp.sYBXZydhn8 + rm /tmp/tmp.l4nFMDf7Bx /tmp/tmp.sYBXZydhn8 + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -ptest-password' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -ptest-password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -ptest-password' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -ptest-password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K5rfsuTVQT +++ mktemp ++ local LAST_ERR=/tmp/tmp.IYcLiF0SoZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.K5rfsuTVQT ++ cat /tmp/tmp.IYcLiF0SoZ ++ rm /tmp/tmp.K5rfsuTVQT /tmp/tmp.IYcLiF0SoZ ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Ri7icm2AFE/select-4.sql + desc 'test proxyadmin' + set +o xtrace ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.fLKfptu57O ++ mktemp + local LAST_ERR=/tmp/tmp.tb1sX1GZdB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.fLKfptu57O perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.tb1sX1GZdB + rm /tmp/tmp.fLKfptu57O /tmp/tmp.tb1sX1GZdB + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yWsoFQYfHO +++ mktemp ++ local LAST_ERR=/tmp/tmp.u7fKUuVYZA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.yWsoFQYfHO ++ cat /tmp/tmp.u7fKUuVYZA ++ rm /tmp/tmp.yWsoFQYfHO /tmp/tmp.u7fKUuVYZA ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vLWUjxPAfj +++ mktemp ++ local LAST_ERR=/tmp/tmp.Xpigqoc5ux ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.vLWUjxPAfj ++ cat /tmp/tmp.Xpigqoc5ux ++ rm /tmp/tmp.vLWUjxPAfj /tmp/tmp.Xpigqoc5ux ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.TPOjGojFTY ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.cgGxfJsFyc +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.TPOjGojFTY +++++ cat /tmp/tmp.cgGxfJsFyc +++++ rm /tmp/tmp.TPOjGojFTY /tmp/tmp.cgGxfJsFyc +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.mfs2v9Pn9c ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.X3WvksjjIe +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.mfs2v9Pn9c +++++ cat /tmp/tmp.X3WvksjjIe +++++ rm /tmp/tmp.mfs2v9Pn9c /tmp/tmp.X3WvksjjIe +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZXIT9hSEqC +++ mktemp ++ local LAST_ERR=/tmp/tmp.gZotebmXHD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.ZXIT9hSEqC ++ cat /tmp/tmp.gZotebmXHD ++ rm /tmp/tmp.ZXIT9hSEqC /tmp/tmp.gZotebmXHD ++ return 0 + [[ 3 == \3 ]] + patch_secret my-cluster-secrets proxyadmin dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=proxyadmin + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.Ot16FrAys4 ++ mktemp + local LAST_ERR=/tmp/tmp.R90MU2tOi2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.Ot16FrAys4 secret/my-cluster-secrets patched + cat /tmp/tmp.R90MU2tOi2 + rm /tmp/tmp.Ot16FrAys4 /tmp/tmp.R90MU2tOi2 + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IFReqR1i40 +++ mktemp ++ local LAST_ERR=/tmp/tmp.whyp2gJC2j ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.IFReqR1i40 ++ cat /tmp/tmp.whyp2gJC2j ++ rm /tmp/tmp.IFReqR1i40 /tmp/tmp.whyp2gJC2j ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HxT2ZX5Bdb +++ mktemp ++ local LAST_ERR=/tmp/tmp.SyhwDXA5ou ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.HxT2ZX5Bdb ++ cat /tmp/tmp.SyhwDXA5ou ++ rm /tmp/tmp.HxT2ZX5Bdb /tmp/tmp.SyhwDXA5ou ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9EXct2UjQw +++ mktemp ++ local LAST_ERR=/tmp/tmp.wvYD7zgokm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.9EXct2UjQw ++ cat /tmp/tmp.wvYD7zgokm ++ rm /tmp/tmp.9EXct2UjQw /tmp/tmp.wvYD7zgokm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TTb0aUppQz +++ mktemp ++ local LAST_ERR=/tmp/tmp.WTBNx2Q7bL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.TTb0aUppQz ++ cat /tmp/tmp.WTBNx2Q7bL ++ rm /tmp/tmp.TTb0aUppQz /tmp/tmp.WTBNx2Q7bL ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rZR5Af9fkm +++ mktemp ++ local LAST_ERR=/tmp/tmp.V1QC5MFoTE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.rZR5Af9fkm ++ cat /tmp/tmp.V1QC5MFoTE ++ rm /tmp/tmp.rZR5Af9fkm /tmp/tmp.V1QC5MFoTE ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Tmutw9fvRH ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.I346j2agLQ +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.Tmutw9fvRH +++++ cat /tmp/tmp.I346j2agLQ +++++ rm /tmp/tmp.Tmutw9fvRH /tmp/tmp.I346j2agLQ +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.hmKYldvWjU ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.E86XcKQ9Fa +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.hmKYldvWjU +++++ cat /tmp/tmp.E86XcKQ9Fa +++++ rm /tmp/tmp.hmKYldvWjU /tmp/tmp.E86XcKQ9Fa +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jQpWtS9DeW +++ mktemp ++ local LAST_ERR=/tmp/tmp.qZqcQc5mqX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.jQpWtS9DeW ++ cat /tmp/tmp.qZqcQc5mqX ++ rm /tmp/tmp.jQpWtS9DeW /tmp/tmp.qZqcQc5mqX ++ return 0 + [[ 3 == \3 ]] + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-0 '' -cproxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-0 + local postfix= + local container_name=-cproxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2-80.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-0 -cproxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-0 + local container_name=-cproxysql + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-2.sql ']' + sleep 20 + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-0 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-0 + local container_name= + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2.sql /tmp/tmp.Ri7icm2AFE/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-1 '' -cproxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-1 + local postfix= + local container_name=-cproxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2-80.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-1 -cproxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-1 + local container_name=-cproxysql + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-2.sql ']' + sleep 20 + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-1 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-1 + local container_name= + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2.sql /tmp/tmp.Ri7icm2AFE/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-2 '' -cproxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-2 + local postfix= + local container_name=-cproxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2-80.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-2 -cproxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-2 + local container_name=-cproxysql + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-2.sql ']' + sleep 20 + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' some-name-proxysql-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -ptest-password' + local pod=some-name-proxysql-2 + local container_name= + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-2.sql /tmp/tmp.Ri7icm2AFE/select-2.sql + desc 'test xtrabackup' + set +o xtrace ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":1}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.yxvjcGCpsL ++ mktemp + local LAST_ERR=/tmp/tmp.mq3Xb4pBSL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":1}}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.yxvjcGCpsL perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.mq3Xb4pBSL + rm /tmp/tmp.yxvjcGCpsL /tmp/tmp.mq3Xb4pBSL + return 0 + patch_secret my-cluster-secrets xtrabackup dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=xtrabackup + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.3xjDo4YtWX ++ mktemp + local LAST_ERR=/tmp/tmp.bGoWhXxS6u + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.3xjDo4YtWX secret/my-cluster-secrets patched + cat /tmp/tmp.bGoWhXxS6u + rm /tmp/tmp.3xjDo4YtWX /tmp/tmp.bGoWhXxS6u + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VAqSULYX7d +++ mktemp ++ local LAST_ERR=/tmp/tmp.sa0BabRmPy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.VAqSULYX7d ++ cat /tmp/tmp.sa0BabRmPy ++ rm /tmp/tmp.VAqSULYX7d /tmp/tmp.sa0BabRmPy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yi78axXnm0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.MWEbToBPaT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.yi78axXnm0 ++ cat /tmp/tmp.MWEbToBPaT ++ rm /tmp/tmp.yi78axXnm0 /tmp/tmp.MWEbToBPaT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JfIQXA1QUK +++ mktemp ++ local LAST_ERR=/tmp/tmp.UHsPyEmi8x ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.JfIQXA1QUK ++ cat /tmp/tmp.UHsPyEmi8x ++ rm /tmp/tmp.JfIQXA1QUK /tmp/tmp.UHsPyEmi8x ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QjaXuiwHmy +++ mktemp ++ local LAST_ERR=/tmp/tmp.evElOUh0Ab ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.QjaXuiwHmy ++ cat /tmp/tmp.evElOUh0Ab ++ rm /tmp/tmp.QjaXuiwHmy /tmp/tmp.evElOUh0Ab ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S3WyyWhEOV +++ mktemp ++ local LAST_ERR=/tmp/tmp.NvQfCjAKRn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.S3WyyWhEOV ++ cat /tmp/tmp.NvQfCjAKRn ++ rm /tmp/tmp.S3WyyWhEOV /tmp/tmp.NvQfCjAKRn ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.YmQwPmKbTR ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Br2tgsmsBM +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.YmQwPmKbTR +++++ cat /tmp/tmp.Br2tgsmsBM +++++ rm /tmp/tmp.YmQwPmKbTR /tmp/tmp.Br2tgsmsBM +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.e0s9sOg7rc ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Lo7X8ubInU +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.e0s9sOg7rc +++++ cat /tmp/tmp.Lo7X8ubInU +++++ rm /tmp/tmp.e0s9sOg7rc /tmp/tmp.Lo7X8ubInU +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5xUsAPnxxO +++ mktemp ++ local LAST_ERR=/tmp/tmp.3NEXdI2Wgn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.5xUsAPnxxO ++ cat /tmp/tmp.3NEXdI2Wgn ++ rm /tmp/tmp.5xUsAPnxxO /tmp/tmp.3NEXdI2Wgn ++ return 0 + [[ 2 == \2 ]] + compare_mysql_cmd_local select-3 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -ptest-password' some-name-pxc-0 '' -cpxc + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -ptest-password' + local pod=some-name-pxc-0 + local postfix= + local container_name=-cpxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-3-80.sql ']' + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -ptest-password' some-name-pxc-0 -cpxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -ptest-password' + local pod=some-name-pxc-0 + local container_name=-cpxc + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-3.sql ']' + sleep 20 + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -ptest-password' some-name-pxc-0 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -ptest-password' + local pod=some-name-pxc-0 + local container_name= + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-3.sql /tmp/tmp.Ri7icm2AFE/select-3.sql + desc 'test clustercheck' + set +o xtrace ----------------------------------------------------------------------------------- test clustercheck ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets clustercheck dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=clustercheck + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"clustercheck": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.d5kLnCkuwI ++ mktemp + local LAST_ERR=/tmp/tmp.yEcYI79RtJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets '-p={"data":{"clustercheck": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.d5kLnCkuwI secret/my-cluster-secrets patched + cat /tmp/tmp.yEcYI79RtJ + rm /tmp/tmp.d5kLnCkuwI /tmp/tmp.yEcYI79RtJ + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eKT3KWBdwc +++ mktemp ++ local LAST_ERR=/tmp/tmp.ISHDIHocVM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.eKT3KWBdwc ++ cat /tmp/tmp.ISHDIHocVM ++ rm /tmp/tmp.eKT3KWBdwc /tmp/tmp.ISHDIHocVM ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ooousb5ync +++ mktemp ++ local LAST_ERR=/tmp/tmp.zCSQzRPG1H ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.ooousb5ync ++ cat /tmp/tmp.zCSQzRPG1H ++ rm /tmp/tmp.ooousb5ync /tmp/tmp.zCSQzRPG1H ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.EDZWeKPdoZ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.O8eWLheeGj +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.EDZWeKPdoZ +++++ cat /tmp/tmp.O8eWLheeGj +++++ rm /tmp/tmp.EDZWeKPdoZ /tmp/tmp.O8eWLheeGj +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.p8XGtYF6wW ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.vFkhnBNUbg +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.p8XGtYF6wW +++++ cat /tmp/tmp.vFkhnBNUbg +++++ rm /tmp/tmp.p8XGtYF6wW /tmp/tmp.vFkhnBNUbg +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5dJcV2WO0h +++ mktemp ++ local LAST_ERR=/tmp/tmp.3yjHYkQN8e ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.5dJcV2WO0h ++ cat /tmp/tmp.3yjHYkQN8e ++ rm /tmp/tmp.5dJcV2WO0h /tmp/tmp.3yjHYkQN8e ++ return 0 + [[ 2 == \2 ]] + compare_mysql_cmd_local select-5 'SHOW DATABASES;' '-h 127.0.0.1 -uclustercheck -ptest-password' some-name-pxc-0 '' -cpxc + local command_id=select-5 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uclustercheck -ptest-password' + local pod=some-name-pxc-0 + local postfix= + local container_name=-cpxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-5.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-5-80.sql ']' + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uclustercheck -ptest-password' some-name-pxc-0 -cpxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uclustercheck -ptest-password' + local pod=some-name-pxc-0 + local container_name=-cpxc + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-5.sql ']' + sleep 20 + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uclustercheck -ptest-password' some-name-pxc-0 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uclustercheck -ptest-password' + local pod=some-name-pxc-0 + local container_name= + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-5.sql /tmp/tmp.Ri7icm2AFE/select-5.sql + desc 'test monitor' + set +o xtrace ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.jFbmUFBE8E ++ mktemp + local LAST_ERR=/tmp/tmp.LM82h6GIK5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.jFbmUFBE8E secret/my-cluster-secrets patched + cat /tmp/tmp.LM82h6GIK5 + rm /tmp/tmp.jFbmUFBE8E /tmp/tmp.LM82h6GIK5 + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iHLemaTKWc +++ mktemp ++ local LAST_ERR=/tmp/tmp.csC3lLOeHp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.iHLemaTKWc ++ cat /tmp/tmp.csC3lLOeHp ++ rm /tmp/tmp.iHLemaTKWc /tmp/tmp.csC3lLOeHp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I8PfNlHmh8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CmLNArx4uR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.I8PfNlHmh8 ++ cat /tmp/tmp.CmLNArx4uR ++ rm /tmp/tmp.I8PfNlHmh8 /tmp/tmp.CmLNArx4uR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NiVx3hnvSo +++ mktemp ++ local LAST_ERR=/tmp/tmp.SiTeQaObZv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.NiVx3hnvSo ++ cat /tmp/tmp.SiTeQaObZv ++ rm /tmp/tmp.NiVx3hnvSo /tmp/tmp.SiTeQaObZv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K8x9vorHjY +++ mktemp ++ local LAST_ERR=/tmp/tmp.QWNwWyXP0w ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.K8x9vorHjY ++ cat /tmp/tmp.QWNwWyXP0w ++ rm /tmp/tmp.K8x9vorHjY /tmp/tmp.QWNwWyXP0w ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zdhCqDEql2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yul2x1Tuar ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.zdhCqDEql2 ++ cat /tmp/tmp.yul2x1Tuar ++ rm /tmp/tmp.zdhCqDEql2 /tmp/tmp.yul2x1Tuar ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.HRosBJOA5G ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.pqKKI6hVXM +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.HRosBJOA5G +++++ cat /tmp/tmp.pqKKI6hVXM +++++ rm /tmp/tmp.HRosBJOA5G /tmp/tmp.pqKKI6hVXM +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.t5N2NOWurk ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Lm2CM5hhwi +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.t5N2NOWurk +++++ cat /tmp/tmp.Lm2CM5hhwi +++++ rm /tmp/tmp.t5N2NOWurk /tmp/tmp.Lm2CM5hhwi +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6uAxt6357b +++ mktemp ++ local LAST_ERR=/tmp/tmp.5WyFNgdEBA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.6uAxt6357b ++ cat /tmp/tmp.5WyFNgdEBA ++ rm /tmp/tmp.6uAxt6357b /tmp/tmp.5WyFNgdEBA ++ return 0 + [[ 2 == \2 ]] + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -ptest-password' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -ptest-password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -ptest-password' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -ptest-password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VtcvDofYXS +++ mktemp ++ local LAST_ERR=/tmp/tmp.pDOWEz9NOJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.VtcvDofYXS ++ cat /tmp/tmp.pDOWEz9NOJ ++ rm /tmp/tmp.VtcvDofYXS /tmp/tmp.pDOWEz9NOJ ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Ri7icm2AFE/select-4.sql + desc 'test operator' + set +o xtrace ----------------------------------------------------------------------------------- test operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets operator dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=operator + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.LwFqqj7xs5 ++ mktemp + local LAST_ERR=/tmp/tmp.avtTzQdeRH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.LwFqqj7xs5 secret/my-cluster-secrets patched + cat /tmp/tmp.avtTzQdeRH + rm /tmp/tmp.LwFqqj7xs5 /tmp/tmp.avtTzQdeRH + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.As4MlZrKLL +++ mktemp ++ local LAST_ERR=/tmp/tmp.HynnGEaTKa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.As4MlZrKLL ++ cat /tmp/tmp.HynnGEaTKa ++ rm /tmp/tmp.As4MlZrKLL /tmp/tmp.HynnGEaTKa ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NLycllW7Sa +++ mktemp ++ local LAST_ERR=/tmp/tmp.0mGph8GTgd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.NLycllW7Sa ++ cat /tmp/tmp.0mGph8GTgd ++ rm /tmp/tmp.NLycllW7Sa /tmp/tmp.0mGph8GTgd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wbBHd3pdLa +++ mktemp ++ local LAST_ERR=/tmp/tmp.HbXhgeRIWJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.wbBHd3pdLa ++ cat /tmp/tmp.HbXhgeRIWJ ++ rm /tmp/tmp.wbBHd3pdLa /tmp/tmp.HbXhgeRIWJ ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uw5rxMzX2g +++ mktemp ++ local LAST_ERR=/tmp/tmp.QeDCnzf2w1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.uw5rxMzX2g ++ cat /tmp/tmp.QeDCnzf2w1 ++ rm /tmp/tmp.uw5rxMzX2g /tmp/tmp.QeDCnzf2w1 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.sUPGtkte1e ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.yfsvOs53Ca +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.sUPGtkte1e +++++ cat /tmp/tmp.yfsvOs53Ca +++++ rm /tmp/tmp.sUPGtkte1e /tmp/tmp.yfsvOs53Ca +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.4wL3xWRbAZ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.yS1sDnCzB7 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.4wL3xWRbAZ +++++ cat /tmp/tmp.yS1sDnCzB7 +++++ rm /tmp/tmp.4wL3xWRbAZ /tmp/tmp.yS1sDnCzB7 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nghtS4xuct +++ mktemp ++ local LAST_ERR=/tmp/tmp.B66tld6Uv6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.nghtS4xuct ++ cat /tmp/tmp.B66tld6Uv6 ++ rm /tmp/tmp.nghtS4xuct /tmp/tmp.B66tld6Uv6 ++ return 0 + [[ 2 == \2 ]] + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -ptest-password' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -ptest-password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -ptest-password' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -ptest-password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LcnbdtaCm8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1Hd4koqTxt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.LcnbdtaCm8 ++ cat /tmp/tmp.1Hd4koqTxt ++ rm /tmp/tmp.LcnbdtaCm8 /tmp/tmp.1Hd4koqTxt ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Ri7icm2AFE/select-4.sql + desc 'change secret name' + set +o xtrace ----------------------------------------------------------------------------------- change secret name ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.WofHT6e2db ++ mktemp + local LAST_ERR=/tmp/tmp.9Ez53n2Ckq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.WofHT6e2db perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.9Ez53n2Ckq + rm /tmp/tmp.WofHT6e2db /tmp/tmp.9Ez53n2Ckq + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gMYPTf0XzR +++ mktemp ++ local LAST_ERR=/tmp/tmp.XtiBK1vMr3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.gMYPTf0XzR ++ cat /tmp/tmp.XtiBK1vMr3 ++ rm /tmp/tmp.gMYPTf0XzR /tmp/tmp.XtiBK1vMr3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KPVcuuXm4H +++ mktemp ++ local LAST_ERR=/tmp/tmp.jW3c7Lgb9o ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.KPVcuuXm4H ++ cat /tmp/tmp.jW3c7Lgb9o ++ rm /tmp/tmp.KPVcuuXm4H /tmp/tmp.jW3c7Lgb9o ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ctQrai8Ef0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.JXYq8NaF2E ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.ctQrai8Ef0 ++ cat /tmp/tmp.JXYq8NaF2E ++ rm /tmp/tmp.ctQrai8Ef0 /tmp/tmp.JXYq8NaF2E ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QbVuLHm3UT +++ mktemp ++ local LAST_ERR=/tmp/tmp.GhYqXs2y5h ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.QbVuLHm3UT ++ cat /tmp/tmp.GhYqXs2y5h ++ rm /tmp/tmp.QbVuLHm3UT /tmp/tmp.GhYqXs2y5h ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vA6pUGjKvw +++ mktemp ++ local LAST_ERR=/tmp/tmp.iZe9FXhsAO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.vA6pUGjKvw ++ cat /tmp/tmp.iZe9FXhsAO ++ rm /tmp/tmp.vA6pUGjKvw /tmp/tmp.iZe9FXhsAO ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Dn0iwZ5xQp ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.bTmdWIeBY0 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.Dn0iwZ5xQp +++++ cat /tmp/tmp.bTmdWIeBY0 +++++ rm /tmp/tmp.Dn0iwZ5xQp /tmp/tmp.bTmdWIeBY0 +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.9N0Zp99MA9 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.VkceiKNv50 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.9N0Zp99MA9 +++++ cat /tmp/tmp.VkceiKNv50 +++++ rm /tmp/tmp.9N0Zp99MA9 /tmp/tmp.VkceiKNv50 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZEv6WIOZqs +++ mktemp ++ local LAST_ERR=/tmp/tmp.ghtNT1QjKb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.ZEv6WIOZqs ++ cat /tmp/tmp.ghtNT1QjKb ++ rm /tmp/tmp.ZEv6WIOZqs /tmp/tmp.ghtNT1QjKb ++ return 0 + [[ 2 == \2 ]] + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + desc 'test new operator' + set +o xtrace ----------------------------------------------------------------------------------- test new operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets-2 operator dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets-2 + local key=operator + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.7CLtIB4jkv ++ mktemp + local LAST_ERR=/tmp/tmp.YeQqEp0TOR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.7CLtIB4jkv secret/my-cluster-secrets-2 patched + cat /tmp/tmp.YeQqEp0TOR + rm /tmp/tmp.7CLtIB4jkv /tmp/tmp.YeQqEp0TOR + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zO42qYqt2f +++ mktemp ++ local LAST_ERR=/tmp/tmp.ADVLNih3So ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.zO42qYqt2f ++ cat /tmp/tmp.ADVLNih3So ++ rm /tmp/tmp.zO42qYqt2f /tmp/tmp.ADVLNih3So ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rvF7jTDa0m +++ mktemp ++ local LAST_ERR=/tmp/tmp.HrUj9R6r0l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.rvF7jTDa0m ++ cat /tmp/tmp.HrUj9R6r0l ++ rm /tmp/tmp.rvF7jTDa0m /tmp/tmp.HrUj9R6r0l ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3hAStEPvq1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.4phkyhUzRz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.3hAStEPvq1 ++ cat /tmp/tmp.4phkyhUzRz ++ rm /tmp/tmp.3hAStEPvq1 /tmp/tmp.4phkyhUzRz ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.633HTRAV5f ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.5ZVX9AU38E +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.633HTRAV5f +++++ cat /tmp/tmp.5ZVX9AU38E +++++ rm /tmp/tmp.633HTRAV5f /tmp/tmp.5ZVX9AU38E +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.aAhu6MCjYv ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.29Tqbvgji1 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.aAhu6MCjYv +++++ cat /tmp/tmp.29Tqbvgji1 +++++ rm /tmp/tmp.aAhu6MCjYv /tmp/tmp.29Tqbvgji1 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kb1SonuNbP +++ mktemp ++ local LAST_ERR=/tmp/tmp.SfuIqiHuvl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.kb1SonuNbP ++ cat /tmp/tmp.SfuIqiHuvl ++ rm /tmp/tmp.kb1SonuNbP /tmp/tmp.SfuIqiHuvl ++ return 0 + [[ 2 == \2 ]] + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -ptest-password2' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -ptest-password2' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -ptest-password2' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -ptest-password2' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pQSpkathQs +++ mktemp ++ local LAST_ERR=/tmp/tmp.luaW4oLEg8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.pQSpkathQs ++ cat /tmp/tmp.luaW4oLEg8 ++ rm /tmp/tmp.pQSpkathQs /tmp/tmp.luaW4oLEg8 ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Ri7icm2AFE/select-4.sql ++ getSecretData my-cluster-secrets-2 root ++ local secretName=my-cluster-secrets-2 ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.YvDkqJWAHQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.B0JiNHZB8n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.YvDkqJWAHQ ++ cat /tmp/tmp.B0JiNHZB8n ++ rm /tmp/tmp.YvDkqJWAHQ /tmp/tmp.B0JiNHZB8n ++ return 0 + newpass=DHmy5IcCvrJFMHBa + desc 'test new users sync' + set +o xtrace ----------------------------------------------------------------------------------- test new users sync ----------------------------------------------------------------------------------- + run_mysql 'CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''DHmy5IcCvrJFMHBa'\'';' '-h some-name-pxc -uroot -pDHmy5IcCvrJFMHBa' + local 'command=CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''DHmy5IcCvrJFMHBa'\'';' + local 'uri=-h some-name-pxc -uroot -pDHmy5IcCvrJFMHBa' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sRkZRmZ6RY +++ mktemp ++ local LAST_ERR=/tmp/tmp.Gz473CfJe9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.sRkZRmZ6RY ++ cat /tmp/tmp.Gz473CfJe9 ++ rm /tmp/tmp.sRkZRmZ6RY /tmp/tmp.Gz473CfJe9 ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + sleep 30 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -utestsync -pDHmy5IcCvrJFMHBa' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -pDHmy5IcCvrJFMHBa' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -utestsync -pDHmy5IcCvrJFMHBa' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -pDHmy5IcCvrJFMHBa' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Cy90p3pl0m +++ mktemp ++ local LAST_ERR=/tmp/tmp.HWzYvBA6yB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.Cy90p3pl0m ++ cat /tmp/tmp.HWzYvBA6yB ++ rm /tmp/tmp.Cy90p3pl0m /tmp/tmp.HWzYvBA6yB ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Ri7icm2AFE/select-4.sql ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.VrxvXCJ8Zs +++ mktemp ++ local LAST_ERR=/tmp/tmp.rz8mdf1QgS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.VrxvXCJ8Zs ++ cat /tmp/tmp.rz8mdf1QgS ++ rm /tmp/tmp.VrxvXCJ8Zs /tmp/tmp.rz8mdf1QgS ++ return 0 + pass=test-password2 + desc 'check secret without operator' + set +o xtrace ----------------------------------------------------------------------------------- check secret without operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.dbBORPmCSj ++ mktemp + local LAST_ERR=/tmp/tmp.skW80ZhUJd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/conf/secrets.yml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.dbBORPmCSj secret/my-cluster-secrets-2 configured + cat /tmp/tmp.skW80ZhUJd Warning: resource secrets/my-cluster-secrets-2 is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.dbBORPmCSj /tmp/tmp.skW80ZhUJd + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -ptest-password2' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -ptest-password2' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -ptest-password2' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -ptest-password2' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GlpVJ7OtgE +++ mktemp ++ local LAST_ERR=/tmp/tmp.H8qkOuJOkR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.GlpVJ7OtgE ++ cat /tmp/tmp.H8qkOuJOkR ++ rm /tmp/tmp.GlpVJ7OtgE /tmp/tmp.H8qkOuJOkR ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Ri7icm2AFE/select-4.sql + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/conf/some-name.yml ++ mktemp + local LAST_OUT=/tmp/tmp.bInwCpkbQh + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1-11-0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#initImage:.*-init$#initImage: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-19289~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/conf/some-name.yml + local LAST_ERR=/tmp/tmp.XWanXAcw4z + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.bInwCpkbQh perconaxtradbcluster.pxc.percona.com/some-name configured + cat /tmp/tmp.XWanXAcw4z + rm /tmp/tmp.bInwCpkbQh /tmp/tmp.XWanXAcw4z + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8EaZ7K1LDN +++ mktemp ++ local LAST_ERR=/tmp/tmp.7Zdm0O0IPd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.8EaZ7K1LDN ++ cat /tmp/tmp.7Zdm0O0IPd ++ rm /tmp/tmp.8EaZ7K1LDN /tmp/tmp.7Zdm0O0IPd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FojdMirCcU +++ mktemp ++ local LAST_ERR=/tmp/tmp.S178YIyONj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.FojdMirCcU ++ cat /tmp/tmp.S178YIyONj ++ rm /tmp/tmp.FojdMirCcU /tmp/tmp.S178YIyONj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x6aEdsTk4R +++ mktemp ++ local LAST_ERR=/tmp/tmp.A70Ncf7Vmo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.x6aEdsTk4R ++ cat /tmp/tmp.A70Ncf7Vmo ++ rm /tmp/tmp.x6aEdsTk4R /tmp/tmp.A70Ncf7Vmo ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hiXDbIgL5c +++ mktemp ++ local LAST_ERR=/tmp/tmp.5Xm09LMU8C ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.hiXDbIgL5c ++ cat /tmp/tmp.5Xm09LMU8C ++ rm /tmp/tmp.hiXDbIgL5c /tmp/tmp.5Xm09LMU8C ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d34eaYvYVw +++ mktemp ++ local LAST_ERR=/tmp/tmp.vKoOQMoaBx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.d34eaYvYVw ++ cat /tmp/tmp.vKoOQMoaBx ++ rm /tmp/tmp.d34eaYvYVw /tmp/tmp.vKoOQMoaBx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Jaog1IhfLD +++ mktemp ++ local LAST_ERR=/tmp/tmp.9JVe1MfInd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.Jaog1IhfLD ++ cat /tmp/tmp.9JVe1MfInd ++ rm /tmp/tmp.Jaog1IhfLD /tmp/tmp.9JVe1MfInd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4Z0iEfUkfP +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z2wsTzwVgR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.4Z0iEfUkfP ++ cat /tmp/tmp.Z2wsTzwVgR ++ rm /tmp/tmp.4Z0iEfUkfP /tmp/tmp.Z2wsTzwVgR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.04vQB2mlp0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.w7tksnmOfo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.04vQB2mlp0 ++ cat /tmp/tmp.w7tksnmOfo ++ rm /tmp/tmp.04vQB2mlp0 /tmp/tmp.w7tksnmOfo ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.b4qKhphrxe +++ mktemp ++ local LAST_ERR=/tmp/tmp.lssAaNPIbx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.b4qKhphrxe ++ cat /tmp/tmp.lssAaNPIbx ++ rm /tmp/tmp.b4qKhphrxe /tmp/tmp.lssAaNPIbx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.M5klwKTEi0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.uJDDRvPAUX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.M5klwKTEi0 ++ cat /tmp/tmp.uJDDRvPAUX ++ rm /tmp/tmp.M5klwKTEi0 /tmp/tmp.uJDDRvPAUX ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8um6QYdiPt +++ mktemp ++ local LAST_ERR=/tmp/tmp.9l6niXX4sY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.8um6QYdiPt ++ cat /tmp/tmp.9l6niXX4sY ++ rm /tmp/tmp.8um6QYdiPt /tmp/tmp.9l6niXX4sY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yhkpOGsS2S +++ mktemp ++ local LAST_ERR=/tmp/tmp.CbLdsZmAmE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.yhkpOGsS2S ++ cat /tmp/tmp.CbLdsZmAmE ++ rm /tmp/tmp.yhkpOGsS2S /tmp/tmp.CbLdsZmAmE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3YhHF5NEDc +++ mktemp ++ local LAST_ERR=/tmp/tmp.t4wcXHBIqk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.3YhHF5NEDc ++ cat /tmp/tmp.t4wcXHBIqk ++ rm /tmp/tmp.3YhHF5NEDc /tmp/tmp.t4wcXHBIqk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9hvB7R8ROd +++ mktemp ++ local LAST_ERR=/tmp/tmp.U72ugbTtON ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.9hvB7R8ROd ++ cat /tmp/tmp.U72ugbTtON ++ rm /tmp/tmp.9hvB7R8ROd /tmp/tmp.U72ugbTtON ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1vJA7PzkXr +++ mktemp ++ local LAST_ERR=/tmp/tmp.QNyZDEAlGh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.1vJA7PzkXr ++ cat /tmp/tmp.QNyZDEAlGh ++ rm /tmp/tmp.1vJA7PzkXr /tmp/tmp.QNyZDEAlGh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.99o2mznZIJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.mG70ZH5fjf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.99o2mznZIJ ++ cat /tmp/tmp.mG70ZH5fjf ++ rm /tmp/tmp.99o2mznZIJ /tmp/tmp.mG70ZH5fjf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0U1L0ZEXjA +++ mktemp ++ local LAST_ERR=/tmp/tmp.Pu9H18On9O ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.0U1L0ZEXjA ++ cat /tmp/tmp.Pu9H18On9O ++ rm /tmp/tmp.0U1L0ZEXjA /tmp/tmp.Pu9H18On9O ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dkZP7CoVs3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.l3YYvCX1PH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.dkZP7CoVs3 ++ cat /tmp/tmp.l3YYvCX1PH ++ rm /tmp/tmp.dkZP7CoVs3 /tmp/tmp.l3YYvCX1PH ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I6737X9y6O +++ mktemp ++ local LAST_ERR=/tmp/tmp.JpCPuAQxez ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.I6737X9y6O ++ cat /tmp/tmp.JpCPuAQxez ++ rm /tmp/tmp.I6737X9y6O /tmp/tmp.JpCPuAQxez ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.B1uM8wd5Hl ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.5zGIIHmEir +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.B1uM8wd5Hl +++++ cat /tmp/tmp.5zGIIHmEir +++++ rm /tmp/tmp.B1uM8wd5Hl /tmp/tmp.5zGIIHmEir +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OYSUv4SPZO +++ mktemp ++ local LAST_ERR=/tmp/tmp.dnmVxTVj6J ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.OYSUv4SPZO ++ cat /tmp/tmp.dnmVxTVj6J ++ rm /tmp/tmp.OYSUv4SPZO /tmp/tmp.dnmVxTVj6J ++ return 0 + [[ 3 == \3 ]] + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.nuqM3w5A3h ++ mktemp + local LAST_ERR=/tmp/tmp.gyNgIbXpNC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.nuqM3w5A3h secret/my-cluster-secrets patched + cat /tmp/tmp.gyNgIbXpNC + rm /tmp/tmp.nuqM3w5A3h /tmp/tmp.gyNgIbXpNC + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1ekAy8isLt +++ mktemp ++ local LAST_ERR=/tmp/tmp.KWzJMM2p41 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.1ekAy8isLt ++ cat /tmp/tmp.KWzJMM2p41 ++ rm /tmp/tmp.1ekAy8isLt /tmp/tmp.KWzJMM2p41 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Tx0rBqYCP1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tGS1Iuoo7p ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.Tx0rBqYCP1 ++ cat /tmp/tmp.tGS1Iuoo7p ++ rm /tmp/tmp.Tx0rBqYCP1 /tmp/tmp.tGS1Iuoo7p ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OPsTUYnibK +++ mktemp ++ local LAST_ERR=/tmp/tmp.EKP2NbAWI3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.OPsTUYnibK ++ cat /tmp/tmp.EKP2NbAWI3 ++ rm /tmp/tmp.OPsTUYnibK /tmp/tmp.EKP2NbAWI3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h20c6NeVMY +++ mktemp ++ local LAST_ERR=/tmp/tmp.WTlddYk1hm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.h20c6NeVMY ++ cat /tmp/tmp.WTlddYk1hm ++ rm /tmp/tmp.h20c6NeVMY /tmp/tmp.WTlddYk1hm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Peo8vrVFee +++ mktemp ++ local LAST_ERR=/tmp/tmp.aKdzVFJwLg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.Peo8vrVFee ++ cat /tmp/tmp.aKdzVFJwLg ++ rm /tmp/tmp.Peo8vrVFee /tmp/tmp.aKdzVFJwLg ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GPNVD9jEXp +++ mktemp ++ local LAST_ERR=/tmp/tmp.hqxG3KMaFV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.GPNVD9jEXp ++ cat /tmp/tmp.hqxG3KMaFV ++ rm /tmp/tmp.GPNVD9jEXp /tmp/tmp.hqxG3KMaFV ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Wavhj9TzC0 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.7EqZSNaI9Q +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ [[ 0 != 0 ]] +++++ break +++++ cat /tmp/tmp.Wavhj9TzC0 +++++ cat /tmp/tmp.7EqZSNaI9Q +++++ rm /tmp/tmp.Wavhj9TzC0 /tmp/tmp.7EqZSNaI9Q +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.91VEUT2sEO +++ mktemp ++ local LAST_ERR=/tmp/tmp.GTrw8jrvKz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.91VEUT2sEO ++ cat /tmp/tmp.GTrw8jrvKz ++ rm /tmp/tmp.91VEUT2sEO /tmp/tmp.GTrw8jrvKz ++ return 0 + [[ 3 == \3 ]] + compare_mysql_cmd select-3 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -ptest-password2' + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -ptest-password2' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-3-80.sql ']' + run_mysql 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -ptest-password2' + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -ptest-password2' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c5mzOIypr3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.IJBIN2yCcK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.c5mzOIypr3 ++ cat /tmp/tmp.IJBIN2yCcK ++ rm /tmp/tmp.c5mzOIypr3 /tmp/tmp.IJBIN2yCcK ++ return 0 + client_pod=pxc-client-5d749ff8b6-bdr9g + wait_pod pxc-client-5d749ff8b6-bdr9g + local pod=pxc-client-5d749ff8b6-bdr9g + local max_retry=480 + local ns= ++ echo pxc-client-5d749ff8b6-bdr9g ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5d749ff8b6-bdr9g.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Ri7icm2AFE/select-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/users/compare/select-3.sql /tmp/tmp.Ri7icm2AFE/select-3.sql + destroy users-19289 + local namespace=users-19289 + local ignore_logs=false + [[ false == \f\a\l\s\e ]] + grep -v 'get backup status: Job.batch' + grep -v 'the object has been modified' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + sort -u + tee /tmp/tmp.Ri7icm2AFE/operator.log + grep -v level=info ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.szD9CliXUg +++ mktemp ++ local LAST_ERR=/tmp/tmp.doA81DsMNK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.szD9CliXUg ++ cat /tmp/tmp.doA81DsMNK ++ rm /tmp/tmp.szD9CliXUg /tmp/tmp.doA81DsMNK ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-5699d7755d-kpc6q ++ mktemp + local LAST_OUT=/tmp/tmp.sl8AYbB3z6 ++ mktemp + local LAST_ERR=/tmp/tmp.kZZ2OPDGcE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-5699d7755d-kpc6q + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.sl8AYbB3z6 + cat /tmp/tmp.kZZ2OPDGcE + rm /tmp/tmp.sl8AYbB3z6 /tmp/tmp.kZZ2OPDGcE + return 0 I0412 12:48:52.472899 1 request.go:645] Throttling request took 1.03430718s, request: GET:https://10.7.240.1:443/apis/pxc.percona.com/v1-10-0?timeout=32s {"level":"error",,"caller":"pxc/controller.go:1142","msg":"sync users","error":"exec syncusers: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-19289.svc.' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-19289.svc.' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-19289.svc.' (using password: YES)\nERROR (line:1921) : Could not find a primary cluster node\n","errorVerbose":"exec syncusers: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-19289.svc.' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-19289.svc.' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-19289.svc.' (using password: YES)\nERROR (line:1921) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:491\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1140\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1581","stacktrace":"github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1142"} {"level":"error",,"caller":"pxc/controller.go:1142","msg":"sync users","error":"exec syncusers: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:581) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n","errorVerbose":"exec syncusers: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:581) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:491\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1140\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1581","stacktrace":"github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1142"} {"level":"error",,"caller":"pxc/controller.go:1142","msg":"sync users","error":"exec syncusers: command terminated with exit code 1 / / ERROR (line:633) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\n","errorVerbose":"exec syncusers: command terminated with exit code 1 / / ERROR (line:633) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:491\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1140\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1581","stacktrace":"github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1142"} {"level":"error",,"caller":"pxc/controller.go:1142","msg":"sync users","error":"exec syncusers: pod some-name-proxysql-2 does not have a host assigned / / ","errorVerbose":"exec syncusers: pod some-name-proxysql-2 does not have a host assigned / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:491\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1140\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1581","stacktrace":"github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1142"} {"level":"error",,"caller":"pxc/controller.go:1142","msg":"sync users","error":"exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / ","errorVerbose":"exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:491\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1140\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1581","stacktrace":"github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1142"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"failed to connect to pod some-name-pxc-0: dial tcp 10.4.1.43:33062: connect: connection refused"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-19289 on 10.7.240.10:53: no such host"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"failed to ensure cluster readonly status: connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-19289 on 10.7.240.10:53: no such host"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.users-19289 on 10.7.240.10:53: no such host"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp 10.4.0.62:33062: connect: connection refused"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp 10.4.0.64:33062: connect: connection refused"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp: lookup some-name-pxc-2.some-name-pxc.users-19289 on 10.7.240.10:53: no such host"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"get primary pxc pod: failed to get proxySQL db: dial tcp 10.4.2.47:6032: connect: no route to host"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"get primary pxc pod: failed to get proxySQL db: dial tcp 10.4.2.48:6032: connect: no route to host"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"get primary pxc pod: failed to get proxySQL db: dial tcp 10.7.254.53:3306: connect: connection refused"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"get primary pxc pod: failed to get proxySQL db: driver: bad connection"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"get primary pxc pod: not found"} {"level":"info",,"caller":"pxc/replication.go:145","msg":"Unable to find primary pod for replication. No pod with name or ip like this","primary name":"some-name-pxc-0.some-name-pxc.users-19289.svc.cluster.local"} {"level":"info",,"caller":"pxc/version.go:328","msg":"update PXC version (fetched from db)","new version":"8.0.27-18.1"} {"level":"info",,"caller":"v1/pxc_types.go:874","msg":"ProxySQL size will be changed from 1 to 2 due to safe config"} {"level":"info",,"caller":"v1/pxc_types.go:875","msg":"Set allowUnsafeConfigurations=true to disable safe configuration"} {"level":"info",,"logger":"cmd","msg":"Git commit: 706f792ae47c369cb3556faff186b6873a8a247f Git branch: PR-1125-706f792a Build time: 2022-04-12T09:09:41Z"} {"level":"info",,"logger":"cmd","msg":"Go OS/Arch: linux/amd64"} {"level":"info",,"logger":"cmd","msg":"Go Version: go1.17.8"} {"level":"info",,"logger":"cmd","msg":"operator-sdk Version: v0.19.4"} {"level":"info",,"logger":"cmd","msg":"Registering Components."} {"level":"info",,"logger":"cmd","msg":"Runs on","platform":"kubernetes","version":"v1.20.15-gke.4100"} {"level":"info",,"logger":"cmd","msg":"Starting the Cmd."} {"level":"info",,"logger":"controller-runtime.certwatcher","msg":"Starting certificate watcher"} {"level":"info",,"logger":"controller-runtime.certwatcher","msg":"Updated current TLS certificate"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterbackup-controller","msg":"Starting Controller"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterbackup-controller","msg":"Starting EventSource","source":"kind source: /, Kind="} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterbackup-controller","msg":"Starting workers","worker count":1} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbcluster-controller","msg":"Starting Controller"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbcluster-controller","msg":"Starting EventSource","source":"kind source: /, Kind="} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbcluster-controller","msg":"Starting workers","worker count":1} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterrestore-controller","msg":"Starting Controller"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterrestore-controller","msg":"Starting EventSource","source":"kind source: /, Kind="} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterrestore-controller","msg":"Starting workers","worker count":1} {"level":"info",,"logger":"controller-runtime.manager","msg":"starting metrics server","path":"/metrics"} {"level":"info",,"logger":"controller-runtime.metrics","msg":"metrics server is starting to listen","addr":":8080"} {"level":"info",,"logger":"controller-runtime.webhook","msg":"registering webhook","path":"/validate-percona-xtradbcluster"} {"level":"info",,"logger":"controller-runtime.webhook","msg":"serving webhook server","host":"","port":9443} {"level":"info",,"logger":"controller-runtime.webhook.webhooks","msg":"starting webhook server"} {"level":"info",,"logger":"leader","msg":"Became the leader."} {"level":"info",,"logger":"leader","msg":"No pre-existing lock was found."} {"level":"info",,"logger":"leader","msg":"Trying to become the leader."} [mysql] 2022/04/12 13:01:25 packets.go:36: read tcp 10.4.0.59:43104->10.4.1.43:33062: read: connection reset by peer [mysql] 2022/04/12 13:02:44 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:04:08 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:07:56 packets.go:36: read tcp 10.4.0.59:40394->10.4.0.65:33062: read: connection reset by peer [mysql] 2022/04/12 13:13:01 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:13:11 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:13:22 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:13:23 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:14:45 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:14:46 packets.go:36: unexpected EOF [mysql] 2022/04/12 13:14:47 packets.go:36: unexpected EOF + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch pxc -n users-19289 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.aOTx0lXL0a ++ mktemp + local LAST_ERR=/tmp/tmp.KkyKKqkssX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.aOTx0lXL0a perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.KkyKKqkssX + rm /tmp/tmp.aOTx0lXL0a /tmp/tmp.KkyKKqkssX + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.SEAWdTiDQ0 ++ mktemp + local LAST_ERR=/tmp/tmp.KKI1PbS26N + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.SEAWdTiDQ0 No resources found + cat /tmp/tmp.KKI1PbS26N + rm /tmp/tmp.SEAWdTiDQ0 /tmp/tmp.KKI1PbS26N + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.RZR7ARVn6D ++ mktemp + local LAST_ERR=/tmp/tmp.sDi6T2iV0R + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.RZR7ARVn6D No resources found + cat /tmp/tmp.sDi6T2iV0R + rm /tmp/tmp.RZR7ARVn6D /tmp/tmp.sDi6T2iV0R + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.mQzFXD6sXu ++ mktemp + local LAST_ERR=/tmp/tmp.6HfbqEitzB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.mQzFXD6sXu validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.6HfbqEitzB + rm /tmp/tmp.mQzFXD6sXu /tmp/tmp.6HfbqEitzB + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.5.4/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace users-19289 + rm -rf /tmp/tmp.Ri7icm2AFE ++ mktemp + local LAST_OUT=/tmp/tmp.m4R1lPIvXi + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.d8xJ3sJRiI ++ mktemp + local LAST_ERR=/tmp/tmp.DFnRkI98EQ + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.185TpROlzp + local exit_status=0 + for i in '$(seq 0 2)' + kubectl delete --grace-period=0 --force=true namespace users-19289 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete --grace-period=0 --force=true namespace pxc-operator