Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/logs/users-8-0.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + create_infra users-13610 + local ns=users-13610 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n users-11047 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.HGkYa8sV2N ++ mktemp + local LAST_ERR=/tmp/tmp.xdj9HRLnEI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HGkYa8sV2N perconaxtradbcluster.pxc.percona.com "some-name" deleted from users-11047 namespace + cat /tmp/tmp.xdj9HRLnEI + rm /tmp/tmp.HGkYa8sV2N /tmp/tmp.xdj9HRLnEI + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.Nw6vvuASbh ++ mktemp + local LAST_ERR=/tmp/tmp.SdnrIwmikW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Nw6vvuASbh No resources found + cat /tmp/tmp.SdnrIwmikW + rm /tmp/tmp.Nw6vvuASbh /tmp/tmp.SdnrIwmikW + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.APNC1KlUi6 ++ mktemp + local LAST_ERR=/tmp/tmp.voRKOdlmli + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.APNC1KlUi6 No resources found + cat /tmp/tmp.voRKOdlmli + rm /tmp/tmp.APNC1KlUi6 /tmp/tmp.voRKOdlmli + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// ++ tail -n1 + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + xargs kubectl delete ns ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.TfUZqwirWX ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.vIsmnsRR8D + local exit_status=0 ++ seq 0 2 + local LAST_OUT=/tmp/tmp.Pu9MTfexit + for i in '$(seq 0 2)' ++ mktemp + set +e + kubectl delete namespace pxc-operator + local LAST_ERR=/tmp/tmp.J5R9w3EW6V + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Pu9MTfexit + cat /tmp/tmp.J5R9w3EW6V + rm /tmp/tmp.Pu9MTfexit /tmp/tmp.J5R9w3EW6V + return 0 namespace "users-11047" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TfUZqwirWX namespace "pxc-operator" deleted + cat /tmp/tmp.vIsmnsRR8D + rm /tmp/tmp.TfUZqwirWX /tmp/tmp.vIsmnsRR8D + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.KpJD5JC1BU ++ mktemp + local LAST_ERR=/tmp/tmp.A8ns8oTbNQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KpJD5JC1BU namespace/pxc-operator created + cat /tmp/tmp.A8ns8oTbNQ + rm /tmp/tmp.KpJD5JC1BU /tmp/tmp.A8ns8oTbNQ + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.EPM9cgCetk +++ mktemp ++ local LAST_ERR=/tmp/tmp.EKzEGMDFLq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EPM9cgCetk ++ cat /tmp/tmp.EKzEGMDFLq ++ rm /tmp/tmp.EPM9cgCetk /tmp/tmp.EKzEGMDFLq ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2233-b3e838fc-5-cluster6 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.jVbOfeHC4i ++ mktemp + local LAST_ERR=/tmp/tmp.2BkopmHf1k + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2233-b3e838fc-5-cluster6 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jVbOfeHC4i Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2233-b3e838fc-5-cluster6" modified. + cat /tmp/tmp.2BkopmHf1k + rm /tmp/tmp.jVbOfeHC4i /tmp/tmp.2BkopmHf1k + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.iRonWIvnER ++ mktemp + local LAST_ERR=/tmp/tmp.2VCtF5syO8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.iRonWIvnER customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.2VCtF5syO8 + rm /tmp/tmp.iRonWIvnER /tmp/tmp.2VCtF5syO8 + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.MgCKzH34a9 ++ mktemp + local LAST_ERR=/tmp/tmp.ViwJ4EcISE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.MgCKzH34a9 clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.ViwJ4EcISE + rm /tmp/tmp.MgCKzH34a9 /tmp/tmp.ViwJ4EcISE + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + kubectl_bin apply -f - ++ mktemp + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2233-b3e838fc^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + local LAST_OUT=/tmp/tmp.nHk56E1ULP + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - ++ mktemp + local LAST_ERR=/tmp/tmp.KU3EZfhP7P + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nHk56E1ULP deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.KU3EZfhP7P + rm /tmp/tmp.nHk56E1ULP /tmp/tmp.KU3EZfhP7P + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.BnuzMdQOAH ++ mktemp + local LAST_ERR=/tmp/tmp.wswiZUR3CY + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BnuzMdQOAH pod/percona-xtradb-cluster-operator-b76dffb8c-jglst condition met + cat /tmp/tmp.wswiZUR3CY + rm /tmp/tmp.BnuzMdQOAH /tmp/tmp.wswiZUR3CY + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.AfHHTNjT6Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.aJNz3FkpGw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AfHHTNjT6Z ++ cat /tmp/tmp.aJNz3FkpGw ++ rm /tmp/tmp.AfHHTNjT6Z /tmp/tmp.aJNz3FkpGw ++ return 0 + wait_pod percona-xtradb-cluster-operator-b76dffb8c-jglst 480 pxc-operator + local pod=percona-xtradb-cluster-operator-b76dffb8c-jglst + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-b76dffb8c-jglst ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-b76dffb8c-jglst condition met waiting for pod/percona-xtradb-cluster-operator-b76dffb8c-jglst to become Ready.Ok + sleep 3 + create_namespace users-13610 + local namespace=users-13610 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces users-13610' + set +o xtrace + xargs kubectl delete ns ----------------------------------------------------------------------------------- cleaned up old namespaces users-13610 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace users-13610 + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.ZFUVE6wXQl + local LAST_OUT=/tmp/tmp.9DvAP2nBby ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.jdxzeKkqhN + local exit_status=0 + local LAST_ERR=/tmp/tmp.xEPwImYCV2 + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + awk '{print$1}' + for i in '$(seq 0 2)' + set +e + kubectl get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-13610 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9DvAP2nBby + cat /tmp/tmp.xEPwImYCV2 + rm /tmp/tmp.9DvAP2nBby /tmp/tmp.xEPwImYCV2 + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-13610 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-13610 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.ZFUVE6wXQl + cat /tmp/tmp.jdxzeKkqhN Error from server (NotFound): namespaces "users-13610" not found + rm /tmp/tmp.ZFUVE6wXQl /tmp/tmp.jdxzeKkqhN + return 1 + : + wait_for_delete namespace/users-13610 + local res=namespace/users-13610 + echo -n 'waiting for namespace/users-13610 to be deleted' waiting for namespace/users-13610 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "users-13610" not found + desc 'create namespace users-13610' + set +o xtrace ----------------------------------------------------------------------------------- create namespace users-13610 ----------------------------------------------------------------------------------- + kubectl_bin create namespace users-13610 ++ mktemp + local LAST_OUT=/tmp/tmp.tAkPiGhDVU ++ mktemp + local LAST_ERR=/tmp/tmp.iftpVkpMAH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace users-13610 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tAkPiGhDVU namespace/users-13610 created + cat /tmp/tmp.iftpVkpMAH + rm /tmp/tmp.tAkPiGhDVU /tmp/tmp.iftpVkpMAH + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.8RBstXzBAh +++ mktemp ++ local LAST_ERR=/tmp/tmp.wUmtpadsiu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8RBstXzBAh ++ cat /tmp/tmp.wUmtpadsiu ++ rm /tmp/tmp.8RBstXzBAh /tmp/tmp.wUmtpadsiu ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2233-b3e838fc-5-cluster6 --namespace=users-13610 ++ mktemp + local LAST_OUT=/tmp/tmp.6iVJaIBnS2 ++ mktemp + local LAST_ERR=/tmp/tmp.A7px4qlyki + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2233-b3e838fc-5-cluster6 --namespace=users-13610 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6iVJaIBnS2 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2233-b3e838fc-5-cluster6" modified. + cat /tmp/tmp.A7px4qlyki + rm /tmp/tmp.6iVJaIBnS2 /tmp/tmp.A7px4qlyki + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9qzYcqPIiP ++ mktemp + local LAST_ERR=/tmp/tmp.uLCWIxiTBC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9qzYcqPIiP secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.uLCWIxiTBC + rm /tmp/tmp.9qzYcqPIiP /tmp/tmp.uLCWIxiTBC + return 0 + desc 'create PXC cluster with 1-password secret' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster with 1-password secret ----------------------------------------------------------------------------------- + newpass=test-password ++ echo -n test-password ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZA== + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/some-name.yml '' '' /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/secrets_one_pass.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/secrets_one_pass.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/secrets_one_pass.yml ++ mktemp + local LAST_OUT=/tmp/tmp.CfxpWMti8p ++ mktemp + local LAST_ERR=/tmp/tmp.cYZpIT2eaj + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/secrets_one_pass.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CfxpWMti8p secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.cYZpIT2eaj + rm /tmp/tmp.CfxpWMti8p /tmp/tmp.cYZpIT2eaj + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/client.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.EhLICaC1CY + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2233-b3e838fc#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ mktemp + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-13610~ + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + local LAST_ERR=/tmp/tmp.ZZG9jdBMe0 + local exit_status=0 + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EhLICaC1CY deployment.apps/pxc-client created + cat /tmp/tmp.ZZG9jdBMe0 + rm /tmp/tmp.EhLICaC1CY /tmp/tmp.ZZG9jdBMe0 + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/some-name.yml ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.aBEoSrLbQ4 + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2233-b3e838fc#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ mktemp + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + local LAST_ERR=/tmp/tmp.4tAcX3oo15 + local exit_status=0 + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-13610~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.aBEoSrLbQ4 perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.4tAcX3oo15 + rm /tmp/tmp.aBEoSrLbQ4 /tmp/tmp.4tAcX3oo15 + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tHmfNBMojX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XdnN1J3Qzj +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.tHmfNBMojX +++ cat /tmp/tmp.XdnN1J3Qzj +++ rm /tmp/tmp.tHmfNBMojX /tmp/tmp.XdnN1J3Qzj +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Un3g6I52vm ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bZDN5ZnkL5 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Un3g6I52vm +++ cat /tmp/tmp.bZDN5ZnkL5 +++ rm /tmp/tmp.Un3g6I52vm /tmp/tmp.bZDN5ZnkL5 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-13610 ++ mktemp + local LAST_OUT=/tmp/tmp.4ZmU05zv4w ++ mktemp + local LAST_ERR=/tmp/tmp.wxJOuQTbvb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-13610 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-13610 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-13610 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.4ZmU05zv4w + cat /tmp/tmp.wxJOuQTbvb error: no matching resources found + rm /tmp/tmp.4ZmU05zv4w /tmp/tmp.wxJOuQTbvb + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.yQ9kxSuJgF +++ mktemp ++ local LAST_ERR=/tmp/tmp.yr9JldUxph ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yQ9kxSuJgF ++ cat /tmp/tmp.yr9JldUxph ++ rm /tmp/tmp.yQ9kxSuJgF /tmp/tmp.yr9JldUxph ++ return 0 + local 'root_pass=^#UX5NPsM!V)Zy6N' + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hrREJtas9K +++ mktemp ++ local LAST_ERR=/tmp/tmp.TLWbAOrifh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hrREJtas9K ++ cat /tmp/tmp.TLWbAOrifh ++ rm /tmp/tmp.hrREJtas9K /tmp/tmp.TLWbAOrifh ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Gm9sgw3IoQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.74rQWHJhDa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Gm9sgw3IoQ ++ cat /tmp/tmp.74rQWHJhDa ++ rm /tmp/tmp.Gm9sgw3IoQ /tmp/tmp.74rQWHJhDa ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v5cQMvPD1d +++ mktemp ++ local LAST_ERR=/tmp/tmp.5K8SlTScpx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v5cQMvPD1d ++ cat /tmp/tmp.5K8SlTScpx ++ rm /tmp/tmp.v5cQMvPD1d /tmp/tmp.5K8SlTScpx ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1.sql /tmp/tmp.Yr7SjFKT5T/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h6OGSUZb6S +++ mktemp ++ local LAST_ERR=/tmp/tmp.CUf20nzB3e ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h6OGSUZb6S ++ cat /tmp/tmp.CUf20nzB3e ++ rm /tmp/tmp.h6OGSUZb6S /tmp/tmp.CUf20nzB3e ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1.sql /tmp/tmp.Yr7SjFKT5T/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''^#UX5NPsM!V)Zy6N'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Eu0ExQ3VTk +++ mktemp ++ local LAST_ERR=/tmp/tmp.v7oPuu0of2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Eu0ExQ3VTk ++ cat /tmp/tmp.v7oPuu0of2 ++ rm /tmp/tmp.Eu0ExQ3VTk /tmp/tmp.v7oPuu0of2 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-1.sql /tmp/tmp.Yr7SjFKT5T/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QEktJsKGaT +++ mktemp ++ local LAST_ERR=/tmp/tmp.TpY85ofneg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QEktJsKGaT ++ cat /tmp/tmp.TpY85ofneg Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.QEktJsKGaT /tmp/tmp.TpY85ofneg ++ return 0 + '[' '' ']' + desc 'test missing passwords were created and present in internal secrets' + set +o xtrace ----------------------------------------------------------------------------------- test missing passwords were created and present in internal secrets ----------------------------------------------------------------------------------- + empty_pwds=() + wrong_pwds=() + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking root' Checking root ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.9ehfjVgOwq +++ mktemp ++ local LAST_ERR=/tmp/tmp.4qFemTx1Yx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9ehfjVgOwq ++ cat /tmp/tmp.4qFemTx1Yx ++ rm /tmp/tmp.9ehfjVgOwq /tmp/tmp.4qFemTx1Yx ++ return 0 + secret_pass='^#UX5NPsM!V)Zy6N' ++ getSecretData internal-some-name root ++ local secretName=internal-some-name ++ local dataKey=root ++ kubectl_bin get secrets/internal-some-name '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.B1IXZCVlbw +++ mktemp ++ local LAST_ERR=/tmp/tmp.ErbxCCZDOo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B1IXZCVlbw ++ cat /tmp/tmp.ErbxCCZDOo ++ rm /tmp/tmp.B1IXZCVlbw /tmp/tmp.ErbxCCZDOo ++ return 0 + int_secret_pass='^#UX5NPsM!V)Zy6N' + [[ -z ^#UX5NPsM!V)Zy6N ]] + [[ ^#UX5NPsM!V)Zy6N != \^\#\U\X\5\N\P\s\M\!\V\)\Z\y\6\N ]] + [[ root != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ root ]] + [[ '' =~ root ]] + echo 'Running compare for root' Running compare for root + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''^#UX5NPsM!V)Zy6N'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Tz4AoQjHMT +++ mktemp ++ local LAST_ERR=/tmp/tmp.VfQZ7bMRd5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Tz4AoQjHMT ++ cat /tmp/tmp.VfQZ7bMRd5 ++ rm /tmp/tmp.Tz4AoQjHMT /tmp/tmp.VfQZ7bMRd5 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking xtrabackup' Checking xtrabackup ++ getSecretData my-cluster-secrets xtrabackup ++ local secretName=my-cluster-secrets ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.vp4jzdlRYy +++ mktemp ++ local LAST_ERR=/tmp/tmp.zbH86rqY7l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vp4jzdlRYy ++ cat /tmp/tmp.zbH86rqY7l ++ rm /tmp/tmp.vp4jzdlRYy /tmp/tmp.zbH86rqY7l ++ return 0 + secret_pass=']TEtg$FhtJ9PQ9dE' ++ getSecretData internal-some-name xtrabackup ++ local secretName=internal-some-name ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.4jLQVZoyxl +++ mktemp ++ local LAST_ERR=/tmp/tmp.YBVBiVzwN2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4jLQVZoyxl ++ cat /tmp/tmp.YBVBiVzwN2 ++ rm /tmp/tmp.4jLQVZoyxl /tmp/tmp.YBVBiVzwN2 ++ return 0 + int_secret_pass=']TEtg$FhtJ9PQ9dE' + [[ -z ]TEtg$FhtJ9PQ9dE ]] + [[ ]TEtg$FhtJ9PQ9dE != \]\T\E\t\g\$\F\h\t\J\9\P\Q\9\d\E ]] + [[ xtrabackup != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ xtrabackup ]] + [[ '' =~ xtrabackup ]] + echo 'Running compare for xtrabackup' Running compare for xtrabackup + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\'']TEtg$FhtJ9PQ9dE'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\'']TEtg$FhtJ9PQ9dE'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\'']TEtg$FhtJ9PQ9dE'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\'']TEtg$FhtJ9PQ9dE'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aaIRMI5PPO +++ mktemp ++ local LAST_ERR=/tmp/tmp.RdTPyu8D35 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aaIRMI5PPO ++ cat /tmp/tmp.RdTPyu8D35 ++ rm /tmp/tmp.aaIRMI5PPO /tmp/tmp.RdTPyu8D35 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking monitor' Checking monitor ++ getSecretData my-cluster-secrets monitor ++ local secretName=my-cluster-secrets ++ local dataKey=monitor ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.HHq7q0dawt +++ mktemp ++ local LAST_ERR=/tmp/tmp.12bc63zoP9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HHq7q0dawt ++ cat /tmp/tmp.12bc63zoP9 ++ rm /tmp/tmp.HHq7q0dawt /tmp/tmp.12bc63zoP9 ++ return 0 + secret_pass=monitor_password ++ getSecretData internal-some-name monitor ++ local secretName=internal-some-name ++ local dataKey=monitor ++ kubectl_bin get secrets/internal-some-name '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.GcNrdZPTxR +++ mktemp ++ local LAST_ERR=/tmp/tmp.FNuFpLgsef ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GcNrdZPTxR ++ cat /tmp/tmp.FNuFpLgsef ++ rm /tmp/tmp.GcNrdZPTxR /tmp/tmp.FNuFpLgsef ++ return 0 + int_secret_pass=monitor_password + [[ -z monitor_password ]] + [[ monitor_password != \m\o\n\i\t\o\r\_\p\a\s\s\w\o\r\d ]] + [[ monitor != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ monitor ]] + [[ '' =~ monitor ]] + echo 'Running compare for monitor' Running compare for monitor + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TydOinjCnV +++ mktemp ++ local LAST_ERR=/tmp/tmp.DOgBfDWFn9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TydOinjCnV ++ cat /tmp/tmp.DOgBfDWFn9 ++ rm /tmp/tmp.TydOinjCnV /tmp/tmp.DOgBfDWFn9 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking proxyadmin' Checking proxyadmin ++ getSecretData my-cluster-secrets proxyadmin ++ local secretName=my-cluster-secrets ++ local dataKey=proxyadmin ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.y9PWKY27h2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HK2GIxyvKX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.y9PWKY27h2 ++ cat /tmp/tmp.HK2GIxyvKX ++ rm /tmp/tmp.y9PWKY27h2 /tmp/tmp.HK2GIxyvKX ++ return 0 + secret_pass='mIZ&j(gs*fA$*Wr^EtJ ]] + [[ 83i=uj7r>A$*Wr^EtJ != \8\3\i\=\u\j\7\r\>\A\$\*\W\r\^\E\t\J ]] + [[ replication != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ replication ]] + [[ '' =~ replication ]] + echo 'Running compare for replication' Running compare for replication + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\''83i=uj7r>A$*Wr^EtJ'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\''83i=uj7r>A$*Wr^EtJ'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\''83i=uj7r>A$*Wr^EtJ'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\''83i=uj7r>A$*Wr^EtJ'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f75vVHxlMo +++ mktemp ++ local LAST_ERR=/tmp/tmp.ugGaax1Gr6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f75vVHxlMo ++ cat /tmp/tmp.ugGaax1Gr6 ++ rm /tmp/tmp.f75vVHxlMo /tmp/tmp.ugGaax1Gr6 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql + [[ -n '' ]] + [[ -n '' ]] + desc 'test root' + set +o xtrace ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets root dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=root + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.AmOVPlF0Yv ++ mktemp + local LAST_ERR=/tmp/tmp.TSBZnFwEyt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.AmOVPlF0Yv secret/my-cluster-secrets patched + cat /tmp/tmp.TSBZnFwEyt + rm /tmp/tmp.AmOVPlF0Yv /tmp/tmp.TSBZnFwEyt + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YWqadsBnqw +++ mktemp ++ local LAST_ERR=/tmp/tmp.dnSb6736Ma ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YWqadsBnqw ++ cat /tmp/tmp.dnSb6736Ma ++ rm /tmp/tmp.YWqadsBnqw /tmp/tmp.dnSb6736Ma ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql + desc 'test proxyadmin' + set +o xtrace ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.VAdziH3TQn ++ mktemp + local LAST_ERR=/tmp/tmp.ckYz7OCCmu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VAdziH3TQn perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.ckYz7OCCmu + rm /tmp/tmp.VAdziH3TQn /tmp/tmp.ckYz7OCCmu + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BGDlMmBpB5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.wmGXQ9WLLY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BGDlMmBpB5 ++ cat /tmp/tmp.wmGXQ9WLLY ++ rm /tmp/tmp.BGDlMmBpB5 /tmp/tmp.wmGXQ9WLLY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ccCLqAjDY6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.f7WU2HEssQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ccCLqAjDY6 ++ cat /tmp/tmp.f7WU2HEssQ ++ rm /tmp/tmp.ccCLqAjDY6 /tmp/tmp.f7WU2HEssQ ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cN1xqJL7MI +++ mktemp ++ local LAST_ERR=/tmp/tmp.oxEYxGCtKN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cN1xqJL7MI ++ cat /tmp/tmp.oxEYxGCtKN ++ rm /tmp/tmp.cN1xqJL7MI /tmp/tmp.oxEYxGCtKN ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.9VE5QrnRRq ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.PY74Ogia70 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.9VE5QrnRRq +++++ cat /tmp/tmp.PY74Ogia70 +++++ rm /tmp/tmp.9VE5QrnRRq /tmp/tmp.PY74Ogia70 +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.HiWZrfZEl3 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.HYeYOYmAWi +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.HiWZrfZEl3 +++++ cat /tmp/tmp.HYeYOYmAWi +++++ rm /tmp/tmp.HiWZrfZEl3 /tmp/tmp.HYeYOYmAWi +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DRyubz4hL3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.dFIHG68DoN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DRyubz4hL3 ++ cat /tmp/tmp.dFIHG68DoN ++ rm /tmp/tmp.DRyubz4hL3 /tmp/tmp.dFIHG68DoN ++ return 0 + [[ 3 == \3 ]] + echo + patch_secret my-cluster-secrets proxyadmin dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=proxyadmin + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.98piM4N8mt ++ mktemp + local LAST_ERR=/tmp/tmp.Z5VNokGwno + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.98piM4N8mt secret/my-cluster-secrets patched + cat /tmp/tmp.Z5VNokGwno + rm /tmp/tmp.98piM4N8mt /tmp/tmp.Z5VNokGwno + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1XFqtysFzW +++ mktemp ++ local LAST_ERR=/tmp/tmp.A6hsEbrSDL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1XFqtysFzW ++ cat /tmp/tmp.A6hsEbrSDL ++ rm /tmp/tmp.1XFqtysFzW /tmp/tmp.A6hsEbrSDL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fazZUW6T1n +++ mktemp ++ local LAST_ERR=/tmp/tmp.LKBHAs1sXb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fazZUW6T1n ++ cat /tmp/tmp.LKBHAs1sXb ++ rm /tmp/tmp.fazZUW6T1n /tmp/tmp.LKBHAs1sXb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gDrEXNIrIx +++ mktemp ++ local LAST_ERR=/tmp/tmp.teYoyywgAV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gDrEXNIrIx ++ cat /tmp/tmp.teYoyywgAV ++ rm /tmp/tmp.gDrEXNIrIx /tmp/tmp.teYoyywgAV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NMb4vEJ8gG +++ mktemp ++ local LAST_ERR=/tmp/tmp.0LRU0oesIQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NMb4vEJ8gG ++ cat /tmp/tmp.0LRU0oesIQ ++ rm /tmp/tmp.NMb4vEJ8gG /tmp/tmp.0LRU0oesIQ ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DnVfegIpy9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.wtQjWrDwAr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DnVfegIpy9 ++ cat /tmp/tmp.wtQjWrDwAr ++ rm /tmp/tmp.DnVfegIpy9 /tmp/tmp.wtQjWrDwAr ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.VvMk2eN5zd ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.vrX5dVgABt +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.VvMk2eN5zd +++++ cat /tmp/tmp.vrX5dVgABt +++++ rm /tmp/tmp.VvMk2eN5zd /tmp/tmp.vrX5dVgABt +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.39UzVouNYd ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.3m3erbF77Y +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.39UzVouNYd +++++ cat /tmp/tmp.3m3erbF77Y +++++ rm /tmp/tmp.39UzVouNYd /tmp/tmp.3m3erbF77Y +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r9cUq64xIb +++ mktemp ++ local LAST_ERR=/tmp/tmp.mbPrs24yOS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.r9cUq64xIb ++ cat /tmp/tmp.mbPrs24yOS ++ rm /tmp/tmp.r9cUq64xIb /tmp/tmp.mbPrs24yOS ++ return 0 + [[ 3 == \3 ]] + echo + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2-80.sql ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2.sql /tmp/tmp.Yr7SjFKT5T/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2-80.sql ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2.sql /tmp/tmp.Yr7SjFKT5T/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2-80.sql ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-2.sql /tmp/tmp.Yr7SjFKT5T/select-2.sql + desc 'test xtrabackup' + set +o xtrace ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.ewmSesVxWs ++ mktemp + local LAST_ERR=/tmp/tmp.vJhuuijhTW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ewmSesVxWs perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.vJhuuijhTW + rm /tmp/tmp.ewmSesVxWs /tmp/tmp.vJhuuijhTW + return 0 + patch_secret my-cluster-secrets xtrabackup dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=xtrabackup + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.f73YpqZqfn ++ mktemp + local LAST_ERR=/tmp/tmp.DRIvlcVn0f + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.f73YpqZqfn secret/my-cluster-secrets patched + cat /tmp/tmp.DRIvlcVn0f + rm /tmp/tmp.f73YpqZqfn /tmp/tmp.DRIvlcVn0f + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z9GGCgd14w +++ mktemp ++ local LAST_ERR=/tmp/tmp.fby7tGpltQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z9GGCgd14w ++ cat /tmp/tmp.fby7tGpltQ ++ rm /tmp/tmp.z9GGCgd14w /tmp/tmp.fby7tGpltQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HM69fz3ul6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.2HpOfFU7vN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HM69fz3ul6 ++ cat /tmp/tmp.2HpOfFU7vN ++ rm /tmp/tmp.HM69fz3ul6 /tmp/tmp.2HpOfFU7vN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3Rb1SfMZbt +++ mktemp ++ local LAST_ERR=/tmp/tmp.xFswnvUy4I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3Rb1SfMZbt ++ cat /tmp/tmp.xFswnvUy4I ++ rm /tmp/tmp.3Rb1SfMZbt /tmp/tmp.xFswnvUy4I ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i4HCSQaJsx +++ mktemp ++ local LAST_ERR=/tmp/tmp.TZalTjbVdG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.i4HCSQaJsx ++ cat /tmp/tmp.TZalTjbVdG ++ rm /tmp/tmp.i4HCSQaJsx /tmp/tmp.TZalTjbVdG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IyXi78CeVK +++ mktemp ++ local LAST_ERR=/tmp/tmp.uGlU0MG4qy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IyXi78CeVK ++ cat /tmp/tmp.uGlU0MG4qy ++ rm /tmp/tmp.IyXi78CeVK /tmp/tmp.uGlU0MG4qy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IM7OAI3024 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WyA8wWWray ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IM7OAI3024 ++ cat /tmp/tmp.WyA8wWWray ++ rm /tmp/tmp.IM7OAI3024 /tmp/tmp.WyA8wWWray ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GGm9JVoPNu +++ mktemp ++ local LAST_ERR=/tmp/tmp.rTSzbloN1M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GGm9JVoPNu ++ cat /tmp/tmp.rTSzbloN1M ++ rm /tmp/tmp.GGm9JVoPNu /tmp/tmp.rTSzbloN1M ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FxNnt2QeIK +++ mktemp ++ local LAST_ERR=/tmp/tmp.tXfn0WQ3A7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FxNnt2QeIK ++ cat /tmp/tmp.tXfn0WQ3A7 ++ rm /tmp/tmp.FxNnt2QeIK /tmp/tmp.tXfn0WQ3A7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f0q82MFol3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.grjZtM3XyD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f0q82MFol3 ++ cat /tmp/tmp.grjZtM3XyD ++ rm /tmp/tmp.f0q82MFol3 /tmp/tmp.grjZtM3XyD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fdWrCkJABH +++ mktemp ++ local LAST_ERR=/tmp/tmp.25GEb9qSH6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fdWrCkJABH ++ cat /tmp/tmp.25GEb9qSH6 ++ rm /tmp/tmp.fdWrCkJABH /tmp/tmp.25GEb9qSH6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ptICvhzQCq +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZMOUTSadqC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ptICvhzQCq ++ cat /tmp/tmp.ZMOUTSadqC ++ rm /tmp/tmp.ptICvhzQCq /tmp/tmp.ZMOUTSadqC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z2726Cc3R5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.n6YNsnqzF4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z2726Cc3R5 ++ cat /tmp/tmp.n6YNsnqzF4 ++ rm /tmp/tmp.z2726Cc3R5 /tmp/tmp.n6YNsnqzF4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9D73eIyzt0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.P4dQRTzhiz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9D73eIyzt0 ++ cat /tmp/tmp.P4dQRTzhiz ++ rm /tmp/tmp.9D73eIyzt0 /tmp/tmp.P4dQRTzhiz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TpUMRMmyDO +++ mktemp ++ local LAST_ERR=/tmp/tmp.88eiu5zAWT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TpUMRMmyDO ++ cat /tmp/tmp.88eiu5zAWT ++ rm /tmp/tmp.TpUMRMmyDO /tmp/tmp.88eiu5zAWT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QNPvT0jjaI +++ mktemp ++ local LAST_ERR=/tmp/tmp.02RX023ujH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QNPvT0jjaI ++ cat /tmp/tmp.02RX023ujH ++ rm /tmp/tmp.QNPvT0jjaI /tmp/tmp.02RX023ujH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ku9LUwinZY +++ mktemp ++ local LAST_ERR=/tmp/tmp.zjMnufPEvh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ku9LUwinZY ++ cat /tmp/tmp.zjMnufPEvh ++ rm /tmp/tmp.Ku9LUwinZY /tmp/tmp.zjMnufPEvh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.obSnXToRZp +++ mktemp ++ local LAST_ERR=/tmp/tmp.VsjzeXhWGP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.obSnXToRZp ++ cat /tmp/tmp.VsjzeXhWGP ++ rm /tmp/tmp.obSnXToRZp /tmp/tmp.VsjzeXhWGP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y1dFvRUDyh +++ mktemp ++ local LAST_ERR=/tmp/tmp.dQxxcuumAO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y1dFvRUDyh ++ cat /tmp/tmp.dQxxcuumAO ++ rm /tmp/tmp.Y1dFvRUDyh /tmp/tmp.dQxxcuumAO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nCbcwzoQrj +++ mktemp ++ local LAST_ERR=/tmp/tmp.kPboUV7bff ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nCbcwzoQrj ++ cat /tmp/tmp.kPboUV7bff ++ rm /tmp/tmp.nCbcwzoQrj /tmp/tmp.kPboUV7bff ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3dvy392xUD +++ mktemp ++ local LAST_ERR=/tmp/tmp.PbjuMUACb9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3dvy392xUD ++ cat /tmp/tmp.PbjuMUACb9 ++ rm /tmp/tmp.3dvy392xUD /tmp/tmp.PbjuMUACb9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CwIhuW6i4l +++ mktemp ++ local LAST_ERR=/tmp/tmp.TtuqMUvvk0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CwIhuW6i4l ++ cat /tmp/tmp.TtuqMUvvk0 ++ rm /tmp/tmp.CwIhuW6i4l /tmp/tmp.TtuqMUvvk0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wVow3rtxKY +++ mktemp ++ local LAST_ERR=/tmp/tmp.K1Ym3km0QV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wVow3rtxKY ++ cat /tmp/tmp.K1Ym3km0QV ++ rm /tmp/tmp.wVow3rtxKY /tmp/tmp.K1Ym3km0QV ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A1yCK0VUeD +++ mktemp ++ local LAST_ERR=/tmp/tmp.2SbRXd9bAW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.A1yCK0VUeD ++ cat /tmp/tmp.2SbRXd9bAW ++ rm /tmp/tmp.A1yCK0VUeD /tmp/tmp.2SbRXd9bAW ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.4hRL00TBe6 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.aMSDZNCoLD +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.4hRL00TBe6 +++++ cat /tmp/tmp.aMSDZNCoLD +++++ rm /tmp/tmp.4hRL00TBe6 /tmp/tmp.aMSDZNCoLD +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.ten5GDMq0r ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.T37RZtzfTv +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.ten5GDMq0r +++++ cat /tmp/tmp.T37RZtzfTv +++++ rm /tmp/tmp.ten5GDMq0r /tmp/tmp.T37RZtzfTv +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eGxBD2SPrT +++ mktemp ++ local LAST_ERR=/tmp/tmp.wGxoaW3s2O ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eGxBD2SPrT ++ cat /tmp/tmp.wGxoaW3s2O ++ rm /tmp/tmp.eGxBD2SPrT /tmp/tmp.wGxoaW3s2O ++ return 0 + [[ 2 == \2 ]] + echo + compare_mysql_cmd_local select-3 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 '' pxc + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-3-80.sql ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 pxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-3.sql /tmp/tmp.Yr7SjFKT5T/select-3.sql + desc 'test monitor' + set +o xtrace ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.SSgvr3fVIw ++ mktemp + local LAST_ERR=/tmp/tmp.CKYgMezIZo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SSgvr3fVIw secret/my-cluster-secrets patched + cat /tmp/tmp.CKYgMezIZo + rm /tmp/tmp.SSgvr3fVIw /tmp/tmp.CKYgMezIZo + return 0 + wait_for_password_propagation my-cluster-secrets monitor + local secret=my-cluster-secrets + local user=monitor + local max_retry=240 ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ base64 --decode ++ local LAST_OUT=/tmp/tmp.tr8L4rutQs +++ mktemp ++ local LAST_ERR=/tmp/tmp.JuaWVkQNDI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tr8L4rutQs ++ cat /tmp/tmp.JuaWVkQNDI ++ rm /tmp/tmp.tr8L4rutQs /tmp/tmp.JuaWVkQNDI ++ return 0 + local root_pass=test-password + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + retry=0 + is_password_updated monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep additional_password + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ch7bHvFHCY +++ mktemp ++ local LAST_ERR=/tmp/tmp.NS6M7xlZD2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ch7bHvFHCY ++ cat /tmp/tmp.NS6M7xlZD2 ++ rm /tmp/tmp.ch7bHvFHCY /tmp/tmp.NS6M7xlZD2 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace {"additional_password": "*C1F414D9BAF378B656A849B31F9F8AF3125F558B"} + retry=0 + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + grep NULL + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rDlahnlM0G +++ mktemp ++ local LAST_ERR=/tmp/tmp.AIMRur4XA6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rDlahnlM0G ++ cat /tmp/tmp.AIMRur4XA6 ++ rm /tmp/tmp.rDlahnlM0G /tmp/tmp.AIMRur4XA6 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 1 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.znNvROcQvE +++ mktemp ++ local LAST_ERR=/tmp/tmp.JxKRqCLUbQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.znNvROcQvE ++ cat /tmp/tmp.JxKRqCLUbQ ++ rm /tmp/tmp.znNvROcQvE /tmp/tmp.JxKRqCLUbQ ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 2 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qeR8s0hhdr +++ mktemp ++ local LAST_ERR=/tmp/tmp.xL5sFVHoVd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qeR8s0hhdr ++ cat /tmp/tmp.xL5sFVHoVd ++ rm /tmp/tmp.qeR8s0hhdr /tmp/tmp.xL5sFVHoVd ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 3 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vng2hFuIBR +++ mktemp ++ local LAST_ERR=/tmp/tmp.xR8a2C29Zu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vng2hFuIBR ++ cat /tmp/tmp.xR8a2C29Zu ++ rm /tmp/tmp.vng2hFuIBR /tmp/tmp.xR8a2C29Zu ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 4 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LP8xTEdFDK +++ mktemp ++ local LAST_ERR=/tmp/tmp.3llL9HOKip ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LP8xTEdFDK ++ cat /tmp/tmp.3llL9HOKip ++ rm /tmp/tmp.LP8xTEdFDK /tmp/tmp.3llL9HOKip ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 5 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + grep NULL + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IYwOiwehpi +++ mktemp ++ local LAST_ERR=/tmp/tmp.Byemt0M1p6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IYwOiwehpi ++ cat /tmp/tmp.Byemt0M1p6 ++ rm /tmp/tmp.IYwOiwehpi /tmp/tmp.Byemt0M1p6 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 6 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZJMvlHavKH +++ mktemp + grep NULL ++ local LAST_ERR=/tmp/tmp.0RQs3dfz2J ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZJMvlHavKH ++ cat /tmp/tmp.0RQs3dfz2J ++ rm /tmp/tmp.ZJMvlHavKH /tmp/tmp.0RQs3dfz2J ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 7 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QRG2ipFTnv +++ mktemp ++ local LAST_ERR=/tmp/tmp.dU5wrD8ORJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QRG2ipFTnv ++ cat /tmp/tmp.dU5wrD8ORJ ++ rm /tmp/tmp.QRG2ipFTnv /tmp/tmp.dU5wrD8ORJ ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 8 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ymLKGH6Cq0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.GkR3FTK1WE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ymLKGH6Cq0 ++ cat /tmp/tmp.GkR3FTK1WE ++ rm /tmp/tmp.ymLKGH6Cq0 /tmp/tmp.GkR3FTK1WE ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 9 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gaFcaTXkij +++ mktemp ++ local LAST_ERR=/tmp/tmp.KOzl2nyVA8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gaFcaTXkij ++ cat /tmp/tmp.KOzl2nyVA8 ++ rm /tmp/tmp.gaFcaTXkij /tmp/tmp.KOzl2nyVA8 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 10 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + grep NULL + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.W3MBRJuDhV +++ mktemp ++ local LAST_ERR=/tmp/tmp.PQZkZt8nEa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.W3MBRJuDhV ++ cat /tmp/tmp.PQZkZt8nEa ++ rm /tmp/tmp.W3MBRJuDhV /tmp/tmp.PQZkZt8nEa ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 11 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w8VRDO2KQz +++ mktemp ++ local LAST_ERR=/tmp/tmp.KaaCtdB91x ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.w8VRDO2KQz ++ cat /tmp/tmp.KaaCtdB91x ++ rm /tmp/tmp.w8VRDO2KQz /tmp/tmp.KaaCtdB91x ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 12 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fmcQYgC1WD +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vnw8Q3pYmv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fmcQYgC1WD ++ cat /tmp/tmp.Vnw8Q3pYmv ++ rm /tmp/tmp.fmcQYgC1WD /tmp/tmp.Vnw8Q3pYmv ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 13 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B8GlOOjV3D +++ mktemp ++ local LAST_ERR=/tmp/tmp.uR9za6jKJn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B8GlOOjV3D ++ cat /tmp/tmp.uR9za6jKJn ++ rm /tmp/tmp.B8GlOOjV3D /tmp/tmp.uR9za6jKJn ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 14 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zryUClyeoE +++ mktemp ++ local LAST_ERR=/tmp/tmp.ObGVDRW9gx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zryUClyeoE ++ cat /tmp/tmp.ObGVDRW9gx ++ rm /tmp/tmp.zryUClyeoE /tmp/tmp.ObGVDRW9gx ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 15 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HZ16iBuE36 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZjZOZttKWl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HZ16iBuE36 ++ cat /tmp/tmp.ZjZOZttKWl ++ rm /tmp/tmp.HZ16iBuE36 /tmp/tmp.ZjZOZttKWl ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + echo 'waiting for password propagation' waiting for password propagation + sleep 1 + let retry+=1 + [[ 16 -ge 240 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + grep NULL + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RaSRKNEoME +++ mktemp ++ local LAST_ERR=/tmp/tmp.Xb1ZEaiYEU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RaSRKNEoME ++ cat /tmp/tmp.Xb1ZEaiYEU ++ rm /tmp/tmp.RaSRKNEoME /tmp/tmp.Xb1ZEaiYEU ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace NULL + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KmxiZ6fa2M +++ mktemp ++ local LAST_ERR=/tmp/tmp.81OhIXg8Ry ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KmxiZ6fa2M ++ cat /tmp/tmp.81OhIXg8Ry ++ rm /tmp/tmp.KmxiZ6fa2M /tmp/tmp.81OhIXg8Ry ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VJ838TvPQI +++ mktemp ++ local LAST_ERR=/tmp/tmp.dLGBYkwfXV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VJ838TvPQI ++ cat /tmp/tmp.dLGBYkwfXV ++ rm /tmp/tmp.VJ838TvPQI /tmp/tmp.dLGBYkwfXV ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.HZ5QA95uOo ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.LB0rJO7oFO +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.HZ5QA95uOo +++++ cat /tmp/tmp.LB0rJO7oFO +++++ rm /tmp/tmp.HZ5QA95uOo /tmp/tmp.LB0rJO7oFO +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.DWeE2EIQm5 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.XUDCgIrBu5 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.DWeE2EIQm5 +++++ cat /tmp/tmp.XUDCgIrBu5 +++++ rm /tmp/tmp.DWeE2EIQm5 /tmp/tmp.XUDCgIrBu5 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HekKeGU5UQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.8bwiX8zu2j ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HekKeGU5UQ ++ cat /tmp/tmp.8bwiX8zu2j ++ rm /tmp/tmp.HekKeGU5UQ /tmp/tmp.8bwiX8zu2j ++ return 0 + [[ 2 == \2 ]] + echo + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jd6VrKsdg8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CEVql5KhJ9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jd6VrKsdg8 ++ cat /tmp/tmp.CEVql5KhJ9 ++ rm /tmp/tmp.jd6VrKsdg8 /tmp/tmp.CEVql5KhJ9 ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql + desc 'test operator' + set +o xtrace ----------------------------------------------------------------------------------- test operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets operator dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=operator + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.ISP73IjWoQ ++ mktemp + local LAST_ERR=/tmp/tmp.owq9jWV6Mn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ISP73IjWoQ secret/my-cluster-secrets patched + cat /tmp/tmp.owq9jWV6Mn + rm /tmp/tmp.ISP73IjWoQ /tmp/tmp.owq9jWV6Mn + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ubPeJOBhnO +++ mktemp ++ local LAST_ERR=/tmp/tmp.ipSJPO3NKH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ubPeJOBhnO ++ cat /tmp/tmp.ipSJPO3NKH ++ rm /tmp/tmp.ubPeJOBhnO /tmp/tmp.ipSJPO3NKH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TFr7VIMPS8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y8VLwc57RG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TFr7VIMPS8 ++ cat /tmp/tmp.Y8VLwc57RG ++ rm /tmp/tmp.TFr7VIMPS8 /tmp/tmp.Y8VLwc57RG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9KpWtP4jQD +++ mktemp ++ local LAST_ERR=/tmp/tmp.3GVKBqw50z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9KpWtP4jQD ++ cat /tmp/tmp.3GVKBqw50z ++ rm /tmp/tmp.9KpWtP4jQD /tmp/tmp.3GVKBqw50z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jhmzf8xGVN +++ mktemp ++ local LAST_ERR=/tmp/tmp.9hQxyG4N7Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jhmzf8xGVN ++ cat /tmp/tmp.9hQxyG4N7Y ++ rm /tmp/tmp.jhmzf8xGVN /tmp/tmp.9hQxyG4N7Y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iAzwJ4CM93 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VyXyebSB1P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iAzwJ4CM93 ++ cat /tmp/tmp.VyXyebSB1P ++ rm /tmp/tmp.iAzwJ4CM93 /tmp/tmp.VyXyebSB1P ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yZKianzd3M +++ mktemp ++ local LAST_ERR=/tmp/tmp.Hxr3it6zmM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yZKianzd3M ++ cat /tmp/tmp.Hxr3it6zmM ++ rm /tmp/tmp.yZKianzd3M /tmp/tmp.Hxr3it6zmM ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nmgxHKy4Tx +++ mktemp ++ local LAST_ERR=/tmp/tmp.aSYWoexK1f ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nmgxHKy4Tx ++ cat /tmp/tmp.aSYWoexK1f ++ rm /tmp/tmp.nmgxHKy4Tx /tmp/tmp.aSYWoexK1f ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.QodtWuhllD ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.O0UX3QujDc +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.QodtWuhllD +++++ cat /tmp/tmp.O0UX3QujDc +++++ rm /tmp/tmp.QodtWuhllD /tmp/tmp.O0UX3QujDc +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.kwxMuSMNAk ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.duqikX6uBr +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.kwxMuSMNAk +++++ cat /tmp/tmp.duqikX6uBr +++++ rm /tmp/tmp.kwxMuSMNAk /tmp/tmp.duqikX6uBr +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l8HnWi1TZE +++ mktemp ++ local LAST_ERR=/tmp/tmp.6FpfpveZKh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.l8HnWi1TZE ++ cat /tmp/tmp.6FpfpveZKh ++ rm /tmp/tmp.l8HnWi1TZE /tmp/tmp.6FpfpveZKh ++ return 0 + [[ 2 == \2 ]] + echo + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bz00uBwQzV +++ mktemp ++ local LAST_ERR=/tmp/tmp.4bOvJwezTF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bz00uBwQzV ++ cat /tmp/tmp.4bOvJwezTF ++ rm /tmp/tmp.bz00uBwQzV /tmp/tmp.4bOvJwezTF ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql + desc 'change secret name' + set +o xtrace ----------------------------------------------------------------------------------- change secret name ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.jmnU5rD4b2 ++ mktemp + local LAST_ERR=/tmp/tmp.3Q14WO2jOr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jmnU5rD4b2 perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.3Q14WO2jOr + rm /tmp/tmp.jmnU5rD4b2 /tmp/tmp.3Q14WO2jOr + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cKgJSKIXjq +++ mktemp ++ local LAST_ERR=/tmp/tmp.TjiaIGZLIo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cKgJSKIXjq ++ cat /tmp/tmp.TjiaIGZLIo ++ rm /tmp/tmp.cKgJSKIXjq /tmp/tmp.TjiaIGZLIo ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LPv8duDr8r +++ mktemp ++ local LAST_ERR=/tmp/tmp.AGSQ7dsSuK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LPv8duDr8r ++ cat /tmp/tmp.AGSQ7dsSuK ++ rm /tmp/tmp.LPv8duDr8r /tmp/tmp.AGSQ7dsSuK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k1lX4TbaKO +++ mktemp ++ local LAST_ERR=/tmp/tmp.xHQAMFjAf9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k1lX4TbaKO ++ cat /tmp/tmp.xHQAMFjAf9 ++ rm /tmp/tmp.k1lX4TbaKO /tmp/tmp.xHQAMFjAf9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1EUp6xGat3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9OgLy0tugG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1EUp6xGat3 ++ cat /tmp/tmp.9OgLy0tugG ++ rm /tmp/tmp.1EUp6xGat3 /tmp/tmp.9OgLy0tugG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YEPAEm9HKF +++ mktemp ++ local LAST_ERR=/tmp/tmp.rXjDVPi3IO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YEPAEm9HKF ++ cat /tmp/tmp.rXjDVPi3IO ++ rm /tmp/tmp.YEPAEm9HKF /tmp/tmp.rXjDVPi3IO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6jv0NxDMjM +++ mktemp ++ local LAST_ERR=/tmp/tmp.GTFRRo4BvQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6jv0NxDMjM ++ cat /tmp/tmp.GTFRRo4BvQ ++ rm /tmp/tmp.6jv0NxDMjM /tmp/tmp.GTFRRo4BvQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JIW6eArw6l +++ mktemp ++ local LAST_ERR=/tmp/tmp.26rcP5Le9T ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JIW6eArw6l ++ cat /tmp/tmp.26rcP5Le9T ++ rm /tmp/tmp.JIW6eArw6l /tmp/tmp.26rcP5Le9T ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OFEYerzBZD +++ mktemp ++ local LAST_ERR=/tmp/tmp.t9UyRKzkLk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OFEYerzBZD ++ cat /tmp/tmp.t9UyRKzkLk ++ rm /tmp/tmp.OFEYerzBZD /tmp/tmp.t9UyRKzkLk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kyQIFbvLIY +++ mktemp ++ local LAST_ERR=/tmp/tmp.fqiE8zeoxD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kyQIFbvLIY ++ cat /tmp/tmp.fqiE8zeoxD ++ rm /tmp/tmp.kyQIFbvLIY /tmp/tmp.fqiE8zeoxD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NdHi7YP8GH +++ mktemp ++ local LAST_ERR=/tmp/tmp.qs6LzifHBE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NdHi7YP8GH ++ cat /tmp/tmp.qs6LzifHBE ++ rm /tmp/tmp.NdHi7YP8GH /tmp/tmp.qs6LzifHBE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yHmRW0ujjG +++ mktemp ++ local LAST_ERR=/tmp/tmp.25MM4HITqr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yHmRW0ujjG ++ cat /tmp/tmp.25MM4HITqr ++ rm /tmp/tmp.yHmRW0ujjG /tmp/tmp.25MM4HITqr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9QUTFlIklQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.aVPkJwwZG9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9QUTFlIklQ ++ cat /tmp/tmp.aVPkJwwZG9 ++ rm /tmp/tmp.9QUTFlIklQ /tmp/tmp.aVPkJwwZG9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f0daMjxQsb +++ mktemp ++ local LAST_ERR=/tmp/tmp.PjG7XvpgwW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f0daMjxQsb ++ cat /tmp/tmp.PjG7XvpgwW ++ rm /tmp/tmp.f0daMjxQsb /tmp/tmp.PjG7XvpgwW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q8kpcSxnzt +++ mktemp ++ local LAST_ERR=/tmp/tmp.s0GeV2bvTM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.q8kpcSxnzt ++ cat /tmp/tmp.s0GeV2bvTM ++ rm /tmp/tmp.q8kpcSxnzt /tmp/tmp.s0GeV2bvTM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xFxYXdOBke +++ mktemp ++ local LAST_ERR=/tmp/tmp.H83Ex5l9DW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xFxYXdOBke ++ cat /tmp/tmp.H83Ex5l9DW ++ rm /tmp/tmp.xFxYXdOBke /tmp/tmp.H83Ex5l9DW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ftrc3ZYMAz +++ mktemp ++ local LAST_ERR=/tmp/tmp.LdDuL1oaOV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ftrc3ZYMAz ++ cat /tmp/tmp.LdDuL1oaOV ++ rm /tmp/tmp.Ftrc3ZYMAz /tmp/tmp.LdDuL1oaOV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CrvRFhNdOe +++ mktemp ++ local LAST_ERR=/tmp/tmp.UrcFgVZqPV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CrvRFhNdOe ++ cat /tmp/tmp.UrcFgVZqPV ++ rm /tmp/tmp.CrvRFhNdOe /tmp/tmp.UrcFgVZqPV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Mt1IPshTaP +++ mktemp ++ local LAST_ERR=/tmp/tmp.mr4jwAV8io ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Mt1IPshTaP ++ cat /tmp/tmp.mr4jwAV8io ++ rm /tmp/tmp.Mt1IPshTaP /tmp/tmp.mr4jwAV8io ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JJr0llrapq +++ mktemp ++ local LAST_ERR=/tmp/tmp.gYZpupFSkV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JJr0llrapq ++ cat /tmp/tmp.gYZpupFSkV ++ rm /tmp/tmp.JJr0llrapq /tmp/tmp.gYZpupFSkV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l6vzml8SLr +++ mktemp ++ local LAST_ERR=/tmp/tmp.eicNGjJEJd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.l6vzml8SLr ++ cat /tmp/tmp.eicNGjJEJd ++ rm /tmp/tmp.l6vzml8SLr /tmp/tmp.eicNGjJEJd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sIHYT132Eq +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZsSOPocOjb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sIHYT132Eq ++ cat /tmp/tmp.ZsSOPocOjb ++ rm /tmp/tmp.sIHYT132Eq /tmp/tmp.ZsSOPocOjb ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.glxRTu7xPw +++ mktemp ++ local LAST_ERR=/tmp/tmp.3lOd1mkgBj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.glxRTu7xPw ++ cat /tmp/tmp.3lOd1mkgBj ++ rm /tmp/tmp.glxRTu7xPw /tmp/tmp.3lOd1mkgBj ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.UQnW6pUNgg ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Dt1IaXKZKa +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.UQnW6pUNgg +++++ cat /tmp/tmp.Dt1IaXKZKa +++++ rm /tmp/tmp.UQnW6pUNgg /tmp/tmp.Dt1IaXKZKa +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.AyLPtZfe4c ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.p1eUTMoOKo +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.AyLPtZfe4c +++++ cat /tmp/tmp.p1eUTMoOKo +++++ rm /tmp/tmp.AyLPtZfe4c /tmp/tmp.p1eUTMoOKo +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IbHJ3cyRmr +++ mktemp ++ local LAST_ERR=/tmp/tmp.ra32VKgSmk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IbHJ3cyRmr ++ cat /tmp/tmp.ra32VKgSmk ++ rm /tmp/tmp.IbHJ3cyRmr /tmp/tmp.ra32VKgSmk ++ return 0 + [[ 2 == \2 ]] + echo + desc 'test new operator' + set +o xtrace ----------------------------------------------------------------------------------- test new operator ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + patch_secret my-cluster-secrets-2 operator dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets-2 + local key=operator + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.Ei07BP5x0Y ++ mktemp + local LAST_ERR=/tmp/tmp.PbyKbrmLLY + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Ei07BP5x0Y secret/my-cluster-secrets-2 patched + cat /tmp/tmp.PbyKbrmLLY + rm /tmp/tmp.Ei07BP5x0Y /tmp/tmp.PbyKbrmLLY + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Uj0Alsi315 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vjiq8UPIn0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Uj0Alsi315 ++ cat /tmp/tmp.Vjiq8UPIn0 ++ rm /tmp/tmp.Uj0Alsi315 /tmp/tmp.Vjiq8UPIn0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dYcOO3DQxg +++ mktemp ++ local LAST_ERR=/tmp/tmp.tQ6aShCZqF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dYcOO3DQxg ++ cat /tmp/tmp.tQ6aShCZqF ++ rm /tmp/tmp.dYcOO3DQxg /tmp/tmp.tQ6aShCZqF ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rG6uZPNZZo +++ mktemp ++ local LAST_ERR=/tmp/tmp.HE1VJtmyUp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rG6uZPNZZo ++ cat /tmp/tmp.HE1VJtmyUp ++ rm /tmp/tmp.rG6uZPNZZo /tmp/tmp.HE1VJtmyUp ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.EnSVszDv32 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.w4AMg6pzmf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.EnSVszDv32 +++++ cat /tmp/tmp.w4AMg6pzmf +++++ rm /tmp/tmp.EnSVszDv32 /tmp/tmp.w4AMg6pzmf +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.jcjvWrODij ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.6brCRGsaAa +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.jcjvWrODij +++++ cat /tmp/tmp.6brCRGsaAa +++++ rm /tmp/tmp.jcjvWrODij /tmp/tmp.6brCRGsaAa +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gVNmk7YBhK +++ mktemp ++ local LAST_ERR=/tmp/tmp.axnYRqvQTl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gVNmk7YBhK ++ cat /tmp/tmp.axnYRqvQTl ++ rm /tmp/tmp.gVNmk7YBhK /tmp/tmp.axnYRqvQTl ++ return 0 + [[ 2 == \2 ]] + echo + sleep 20 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5JD693ymwE +++ mktemp ++ local LAST_ERR=/tmp/tmp.fqVshXR2Ay ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5JD693ymwE ++ cat /tmp/tmp.fqVshXR2Ay ++ rm /tmp/tmp.5JD693ymwE /tmp/tmp.fqVshXR2Ay ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql ++ getSecretData my-cluster-secrets-2 root ++ local secretName=my-cluster-secrets-2 ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dg6KwfP9YT +++ mktemp ++ local LAST_ERR=/tmp/tmp.jZPycTo9wx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Dg6KwfP9YT ++ cat /tmp/tmp.jZPycTo9wx ++ rm /tmp/tmp.Dg6KwfP9YT /tmp/tmp.jZPycTo9wx ++ return 0 + newpass='ragg~f$3?V^-C0XB6J' + desc 'test new users sync' + set +o xtrace ----------------------------------------------------------------------------------- test new users sync ----------------------------------------------------------------------------------- + run_mysql 'CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''ragg~f$3?V^-C0XB6J'\'';' '-h some-name-pxc -uroot -p'\''ragg~f$3?V^-C0XB6J'\''' + local 'command=CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''ragg~f$3?V^-C0XB6J'\'';' + local 'uri=-h some-name-pxc -uroot -p'\''ragg~f$3?V^-C0XB6J'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mY9cXQr1n1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.i61WMfIMzB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mY9cXQr1n1 ++ cat /tmp/tmp.i61WMfIMzB ++ rm /tmp/tmp.mY9cXQr1n1 /tmp/tmp.i61WMfIMzB ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 40 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''ragg~f$3?V^-C0XB6J'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''ragg~f$3?V^-C0XB6J'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''ragg~f$3?V^-C0XB6J'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''ragg~f$3?V^-C0XB6J'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yyHG5MPrE0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.shClflbrRN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yyHG5MPrE0 ++ cat /tmp/tmp.shClflbrRN ++ rm /tmp/tmp.yyHG5MPrE0 /tmp/tmp.shClflbrRN ++ return 0 + client_pod=pxc-client-59944c5bbf-xmcmg + wait_pod pxc-client-59944c5bbf-xmcmg + local pod=pxc-client-59944c5bbf-xmcmg + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-xmcmg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-xmcmg condition met waiting for pod/pxc-client-59944c5bbf-xmcmg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Yr7SjFKT5T/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql /tmp/tmp.Yr7SjFKT5T/select-4.sql --- /mnt/jenkins/workspace/cloud-pxc-operator_PR-2233/e2e-tests/users/compare/select-4-80.sql 2025-11-07 12:42:29.749644347 +0000 +++ /tmp/tmp.Yr7SjFKT5T/select-4.sql 2025-11-07 15:22:28.829512317 +0000 @@ -1,88 +1,2 @@ -ADMINISTRABLE_ROLE_AUTHORIZATIONS -APPLICABLE_ROLES -CHARACTER_SETS -CHECK_CONSTRAINTS -CLIENT_STATISTICS -COLLATIONS -COLLATION_CHARACTER_SET_APPLICABILITY -COLUMNS -COLUMNS_EXTENSIONS -COLUMN_PRIVILEGES -COLUMN_STATISTICS -COMPRESSION_DICTIONARY -COMPRESSION_DICTIONARY_TABLES -ENABLED_ROLES -ENGINES -EVENTS -FILES -GLOBAL_TEMPORARY_TABLES -INDEX_STATISTICS -INNODB_BUFFER_PAGE -INNODB_BUFFER_PAGE_LRU -INNODB_BUFFER_POOL_STATS -INNODB_CACHED_INDEXES -INNODB_CMP -INNODB_CMPMEM -INNODB_CMPMEM_RESET -INNODB_CMP_PER_INDEX -INNODB_CMP_PER_INDEX_RESET -INNODB_CMP_RESET -INNODB_COLUMNS -INNODB_DATAFILES -INNODB_FIELDS -INNODB_FOREIGN -INNODB_FOREIGN_COLS -INNODB_FT_BEING_DELETED -INNODB_FT_CONFIG -INNODB_FT_DEFAULT_STOPWORD -INNODB_FT_DELETED -INNODB_FT_INDEX_CACHE -INNODB_FT_INDEX_TABLE -INNODB_INDEXES -INNODB_METRICS -INNODB_SESSION_TEMP_TABLESPACES -INNODB_TABLES -INNODB_TABLESPACES -INNODB_TABLESPACES_BRIEF -INNODB_TABLESTATS -INNODB_TEMP_TABLE_INFO -INNODB_TRX -INNODB_VIRTUAL -KEYWORDS -KEY_COLUMN_USAGE -OPTIMIZER_TRACE -PARAMETERS -PARTITIONS -PLUGINS -PROCESSLIST -PROFILING -REFERENTIAL_CONSTRAINTS -RESOURCE_GROUPS -ROLE_COLUMN_GRANTS -ROLE_ROUTINE_GRANTS -ROLE_TABLE_GRANTS -ROUTINES -SCHEMATA -SCHEMATA_EXTENSIONS -SCHEMA_PRIVILEGES -STATISTICS -ST_GEOMETRY_COLUMNS -ST_SPATIAL_REFERENCE_SYSTEMS -ST_UNITS_OF_MEASURE -TABLES -TABLESPACES -TABLESPACES_EXTENSIONS -TABLES_EXTENSIONS -TABLE_CONSTRAINTS -TABLE_CONSTRAINTS_EXTENSIONS -TABLE_PRIVILEGES -TABLE_STATISTICS -TEMPORARY_TABLES -THREAD_STATISTICS -TRIGGERS -USER_ATTRIBUTES -USER_PRIVILEGES -USER_STATISTICS -VIEWS -VIEW_ROUTINE_USAGE -VIEW_TABLE_USAGE +ERROR 1045 (28000): ProxySQL Error: Access denied for user 'testsync'@'10.15.241.44' (using password: YES) +command terminated with exit code 1