Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/logs/users-5-7.log WARNING: version difference between client (1.30) and server (1.27) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.30) and server (1.27) exceeds the supported minor version skew of +/-1 + create_infra users-24945 + local ns=users-24945 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n users-24515 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.qWADj4pwAr ++ mktemp + local LAST_ERR=/tmp/tmp.qjjs2x8jdq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qWADj4pwAr perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.qjjs2x8jdq + rm /tmp/tmp.qWADj4pwAr /tmp/tmp.qjjs2x8jdq + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.hIfjI6Jf00 ++ mktemp + local LAST_ERR=/tmp/tmp.gkTP5omTuR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hIfjI6Jf00 No resources found + cat /tmp/tmp.gkTP5omTuR + rm /tmp/tmp.hIfjI6Jf00 /tmp/tmp.gkTP5omTuR + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.26mQJP8FgP ++ mktemp + local LAST_ERR=/tmp/tmp.ueaMqUijni + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.26mQJP8FgP No resources found + cat /tmp/tmp.ueaMqUijni + rm /tmp/tmp.26mQJP8FgP /tmp/tmp.ueaMqUijni + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.0l3PwXqaZ7 + local LAST_OUT=/tmp/tmp.F5S1xfWuVy ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.EI4ANRvhbI + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.Hzt78NsaWx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0l3PwXqaZ7 + cat /tmp/tmp.EI4ANRvhbI + rm /tmp/tmp.0l3PwXqaZ7 /tmp/tmp.EI4ANRvhbI + return 0 namespace "gmp-public" deleted namespace "gmp-system" deleted namespace "users-24515" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.F5S1xfWuVy namespace "pxc-operator" deleted + cat /tmp/tmp.Hzt78NsaWx + rm /tmp/tmp.F5S1xfWuVy /tmp/tmp.Hzt78NsaWx + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'namespace/pxc-operator - ' namespace/pxc-operator - + set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.nUjAlH2FIx ++ mktemp + local LAST_ERR=/tmp/tmp.RloY1Qppr2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nUjAlH2FIx namespace/pxc-operator created + cat /tmp/tmp.RloY1Qppr2 + rm /tmp/tmp.nUjAlH2FIx /tmp/tmp.RloY1Qppr2 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.O1v5Q2zAWY +++ mktemp ++ local LAST_ERR=/tmp/tmp.qZLTruHfu4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.O1v5Q2zAWY ++ cat /tmp/tmp.qZLTruHfu4 ++ rm /tmp/tmp.O1v5Q2zAWY /tmp/tmp.qZLTruHfu4 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1781-4641617b-1-cluster3 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.6icFuCDe2z ++ mktemp + local LAST_ERR=/tmp/tmp.BpJ4GmWld1 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1781-4641617b-1-cluster3 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6icFuCDe2z Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1781-4641617b-1-cluster3" modified. + cat /tmp/tmp.BpJ4GmWld1 + rm /tmp/tmp.6icFuCDe2z /tmp/tmp.BpJ4GmWld1 + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.fq3I0r7hjg ++ mktemp + local LAST_ERR=/tmp/tmp.r9rtjVhCWW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fq3I0r7hjg customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.r9rtjVhCWW + rm /tmp/tmp.fq3I0r7hjg /tmp/tmp.r9rtjVhCWW + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/deploy/cw-rbac.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.lWVcw5SAJC ++ mktemp + local LAST_ERR=/tmp/tmp.cK2iuyVlcJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.lWVcw5SAJC clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.cK2iuyVlcJ + rm /tmp/tmp.lWVcw5SAJC /tmp/tmp.cK2iuyVlcJ + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1781-4641617b^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "DEBUG"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.go7uyTbHtQ ++ mktemp + local LAST_ERR=/tmp/tmp.8dkiFt9Ndu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.go7uyTbHtQ deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.8dkiFt9Ndu + rm /tmp/tmp.go7uyTbHtQ /tmp/tmp.8dkiFt9Ndu + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.skJ2RRzjfk ++ mktemp + local LAST_ERR=/tmp/tmp.0tuuBskKOW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.skJ2RRzjfk pod/percona-xtradb-cluster-operator-86c87db468-n8lhv condition met + cat /tmp/tmp.0tuuBskKOW + rm /tmp/tmp.skJ2RRzjfk /tmp/tmp.0tuuBskKOW + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.0EgInAIlE5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.V3OXCRZb64 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0EgInAIlE5 ++ cat /tmp/tmp.V3OXCRZb64 ++ rm /tmp/tmp.0EgInAIlE5 /tmp/tmp.V3OXCRZb64 ++ return 0 + wait_pod percona-xtradb-cluster-operator-86c87db468-n8lhv 480 pxc-operator + local pod=percona-xtradb-cluster-operator-86c87db468-n8lhv + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-86c87db468-n8lhv ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-86c87db468-n8lhv condition met percona-xtradb-cluster-operator-86c87db468-n8lhv.Ok + sleep 3 + create_namespace users-24945 + local namespace=users-24945 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces users-24945' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces users-24945 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace users-24945 + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.JatQLGq7XS + local LAST_OUT=/tmp/tmp.cAtvHql6AR ++ mktemp + local LAST_ERR=/tmp/tmp.WidtjbbpLr + local exit_status=0 ++ mktemp + local LAST_ERR=/tmp/tmp.HutKIdRJH0 + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-24945 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cAtvHql6AR + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-24945 + cat /tmp/tmp.HutKIdRJH0 + rm /tmp/tmp.cAtvHql6AR /tmp/tmp.HutKIdRJH0 + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-24945 namespace "gmp-public" deleted namespace "gmp-system" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.JatQLGq7XS + cat /tmp/tmp.WidtjbbpLr Error from server (NotFound): namespaces "users-24945" not found + rm /tmp/tmp.JatQLGq7XS /tmp/tmp.WidtjbbpLr + return 1 + : + wait_for_delete namespace/users-24945 + local res=namespace/users-24945 + echo -n 'namespace/users-24945 - ' namespace/users-24945 - + set +o xtrace Error from server (NotFound): namespaces "users-24945" not found + desc 'create namespace users-24945' + set +o xtrace ----------------------------------------------------------------------------------- create namespace users-24945 ----------------------------------------------------------------------------------- + kubectl_bin create namespace users-24945 ++ mktemp + local LAST_OUT=/tmp/tmp.TAG8PFMY3X ++ mktemp + local LAST_ERR=/tmp/tmp.7dfbTLh9OW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace users-24945 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TAG8PFMY3X namespace/users-24945 created + cat /tmp/tmp.7dfbTLh9OW + rm /tmp/tmp.TAG8PFMY3X /tmp/tmp.7dfbTLh9OW + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.WcoYqju2NE +++ mktemp ++ local LAST_ERR=/tmp/tmp.IlKEtLZkyG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WcoYqju2NE ++ cat /tmp/tmp.IlKEtLZkyG ++ rm /tmp/tmp.WcoYqju2NE /tmp/tmp.IlKEtLZkyG ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1781-4641617b-1-cluster3 --namespace=users-24945 ++ mktemp + local LAST_OUT=/tmp/tmp.diHRsbVrhX ++ mktemp + local LAST_ERR=/tmp/tmp.Ipu79AwUtd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1781-4641617b-1-cluster3 --namespace=users-24945 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.diHRsbVrhX Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1781-4641617b-1-cluster3" modified. + cat /tmp/tmp.Ipu79AwUtd + rm /tmp/tmp.diHRsbVrhX /tmp/tmp.Ipu79AwUtd + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.zBGMsz5W1M ++ mktemp + local LAST_ERR=/tmp/tmp.xK2Q2Ktru5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zBGMsz5W1M secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.xK2Q2Ktru5 + rm /tmp/tmp.zBGMsz5W1M /tmp/tmp.xK2Q2Ktru5 + return 0 + desc 'create PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- + newpass=test-password ++ echo -n test-password ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZA== + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.0qrm5ddQuG ++ mktemp + local LAST_ERR=/tmp/tmp.jSRDEkwhL8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0qrm5ddQuG secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.jSRDEkwhL8 + rm /tmp/tmp.0qrm5ddQuG /tmp/tmp.jSRDEkwhL8 + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' ++ mktemp + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1781-4641617b#' + local LAST_OUT=/tmp/tmp.9DFWTgeGeT + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-24945~ ++ mktemp + /usr/bin/sed -e 's#apply:.*#apply: Never#' + local LAST_ERR=/tmp/tmp.dRzg0YedfN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9DFWTgeGeT deployment.apps/pxc-client created + cat /tmp/tmp.dRzg0YedfN + rm /tmp/tmp.9DFWTgeGeT /tmp/tmp.dRzg0YedfN + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p\-\c\l\o\u\d ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/some-name.yml ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + local LAST_OUT=/tmp/tmp.xfYFp0YO1s + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1781-4641617b#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' ++ mktemp + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + local LAST_ERR=/tmp/tmp.W1tmAVQol7 + local exit_status=0 + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-24945~ ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xfYFp0YO1s perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.W1tmAVQol7 + rm /tmp/tmp.xfYFp0YO1s /tmp/tmp.W1tmAVQol7 + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VpLEhpNXH9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.xuSsJk5K0Q +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.VpLEhpNXH9 +++ cat /tmp/tmp.xuSsJk5K0Q +++ rm /tmp/tmp.VpLEhpNXH9 /tmp/tmp.xuSsJk5K0Q +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Gxq1aMnEZb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.6qrLr5536A +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Gxq1aMnEZb +++ cat /tmp/tmp.6qrLr5536A +++ rm /tmp/tmp.Gxq1aMnEZb /tmp/tmp.6qrLr5536A +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-24945 ++ mktemp + local LAST_OUT=/tmp/tmp.6bS5X8v74D ++ mktemp + local LAST_ERR=/tmp/tmp.rlxwJQmVDl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-24945 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-24945 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-24945 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.6bS5X8v74D + cat /tmp/tmp.rlxwJQmVDl error: no matching resources found + rm /tmp/tmp.6bS5X8v74D /tmp/tmp.rlxwJQmVDl + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met some-name-proxysql-0.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met some-name-pxc-0.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met some-name-pxc-1.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met some-name-pxc-2.Ok + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3O53aeS9aR +++ mktemp ++ local LAST_ERR=/tmp/tmp.QqXWPTQtZZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3O53aeS9aR ++ cat /tmp/tmp.QqXWPTQtZZ Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.3O53aeS9aR /tmp/tmp.QqXWPTQtZZ ++ return 0 + [[ -n '' ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -proot_password -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4T36rK4UPo +++ mktemp ++ local LAST_ERR=/tmp/tmp.GHyxguqVUR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4T36rK4UPo ++ cat /tmp/tmp.GHyxguqVUR ++ rm /tmp/tmp.4T36rK4UPo /tmp/tmp.GHyxguqVUR ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -proot_password -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6Rl0uvi9W2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jE8akCCdB5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6Rl0uvi9W2 ++ cat /tmp/tmp.jE8akCCdB5 ++ rm /tmp/tmp.6Rl0uvi9W2 /tmp/tmp.jE8akCCdB5 ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KDbnAzOCAF +++ mktemp ++ local LAST_ERR=/tmp/tmp.MKT0djJjv2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KDbnAzOCAF ++ cat /tmp/tmp.MKT0djJjv2 ++ rm /tmp/tmp.KDbnAzOCAF /tmp/tmp.MKT0djJjv2 ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-1.sql /tmp/tmp.1rIt3CjLpl/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wcFP2ruDVT +++ mktemp ++ local LAST_ERR=/tmp/tmp.Fzsmk3kZIg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wcFP2ruDVT ++ cat /tmp/tmp.Fzsmk3kZIg ++ rm /tmp/tmp.wcFP2ruDVT /tmp/tmp.Fzsmk3kZIg ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-1.sql /tmp/tmp.1rIt3CjLpl/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vZaElMZk2c +++ mktemp ++ local LAST_ERR=/tmp/tmp.5DWGaPPaXJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vZaElMZk2c ++ cat /tmp/tmp.5DWGaPPaXJ ++ rm /tmp/tmp.vZaElMZk2c /tmp/tmp.5DWGaPPaXJ ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-1.sql /tmp/tmp.1rIt3CjLpl/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kzomNVHj5G +++ mktemp ++ local LAST_ERR=/tmp/tmp.s9N8rjvRAs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kzomNVHj5G ++ cat /tmp/tmp.s9N8rjvRAs Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.kzomNVHj5G /tmp/tmp.s9N8rjvRAs ++ return 0 + '[' '' ']' + desc 'test root' + set +o xtrace ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets root dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=root + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.EW0bJ9PQhB ++ mktemp + local LAST_ERR=/tmp/tmp.kmHsFBUdM1 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EW0bJ9PQhB secret/my-cluster-secrets patched + cat /tmp/tmp.kmHsFBUdM1 + rm /tmp/tmp.EW0bJ9PQhB /tmp/tmp.kmHsFBUdM1 + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NtYGMeo1BS +++ mktemp ++ local LAST_ERR=/tmp/tmp.bm5nKm4fbM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NtYGMeo1BS ++ cat /tmp/tmp.bm5nKm4fbM ++ rm /tmp/tmp.NtYGMeo1BS /tmp/tmp.bm5nKm4fbM ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql /tmp/tmp.1rIt3CjLpl/select-4.sql + desc 'test proxyadmin' + set +o xtrace ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.8zU82GqC4n ++ mktemp + local LAST_ERR=/tmp/tmp.FxRhuCPuoE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.8zU82GqC4n perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.FxRhuCPuoE + rm /tmp/tmp.8zU82GqC4n /tmp/tmp.FxRhuCPuoE + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oXO5jFI2cU +++ mktemp ++ local LAST_ERR=/tmp/tmp.7NYmK4N8Y1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oXO5jFI2cU ++ cat /tmp/tmp.7NYmK4N8Y1 ++ rm /tmp/tmp.oXO5jFI2cU /tmp/tmp.7NYmK4N8Y1 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xhjFSgzzT4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3Hw5GgDkga ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xhjFSgzzT4 ++ cat /tmp/tmp.3Hw5GgDkga ++ rm /tmp/tmp.xhjFSgzzT4 /tmp/tmp.3Hw5GgDkga ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.29s7Cun0A4 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.YQ4rpkL0If +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.29s7Cun0A4 +++++ cat /tmp/tmp.YQ4rpkL0If +++++ rm /tmp/tmp.29s7Cun0A4 /tmp/tmp.YQ4rpkL0If +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.WSPrcID46Y ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.eXAuZClQS0 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.WSPrcID46Y +++++ cat /tmp/tmp.eXAuZClQS0 +++++ rm /tmp/tmp.WSPrcID46Y /tmp/tmp.eXAuZClQS0 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FqZPxXharY +++ mktemp ++ local LAST_ERR=/tmp/tmp.avMvsSPJGT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FqZPxXharY ++ cat /tmp/tmp.avMvsSPJGT ++ rm /tmp/tmp.FqZPxXharY /tmp/tmp.avMvsSPJGT ++ return 0 + [[ 3 == \3 ]] + patch_secret my-cluster-secrets proxyadmin dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=proxyadmin + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.UFaSMOt8zT ++ mktemp + local LAST_ERR=/tmp/tmp.7wei6Rf1GZ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.UFaSMOt8zT secret/my-cluster-secrets patched + cat /tmp/tmp.7wei6Rf1GZ + rm /tmp/tmp.UFaSMOt8zT /tmp/tmp.7wei6Rf1GZ + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8jBcM7xxfw +++ mktemp ++ local LAST_ERR=/tmp/tmp.c4xa1Lp3Ew ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8jBcM7xxfw ++ cat /tmp/tmp.c4xa1Lp3Ew ++ rm /tmp/tmp.8jBcM7xxfw /tmp/tmp.c4xa1Lp3Ew ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.W1QoX4KmSA +++ mktemp ++ local LAST_ERR=/tmp/tmp.4j1OfpsOVb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.W1QoX4KmSA ++ cat /tmp/tmp.4j1OfpsOVb ++ rm /tmp/tmp.W1QoX4KmSA /tmp/tmp.4j1OfpsOVb ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Mvg6ULFVeo +++ mktemp ++ local LAST_ERR=/tmp/tmp.SSFRM5RHIn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Mvg6ULFVeo ++ cat /tmp/tmp.SSFRM5RHIn ++ rm /tmp/tmp.Mvg6ULFVeo /tmp/tmp.SSFRM5RHIn ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.LOi3kSPmiL ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.dyqqKaFFa7 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.LOi3kSPmiL +++++ cat /tmp/tmp.dyqqKaFFa7 +++++ rm /tmp/tmp.LOi3kSPmiL /tmp/tmp.dyqqKaFFa7 +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.8rtiNWhEI9 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.ZIojJiTTHO +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.8rtiNWhEI9 +++++ cat /tmp/tmp.ZIojJiTTHO +++++ rm /tmp/tmp.8rtiNWhEI9 /tmp/tmp.ZIojJiTTHO +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dpf1UXaOt8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.6SkOZtwIv8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dpf1UXaOt8 ++ cat /tmp/tmp.6SkOZtwIv8 ++ rm /tmp/tmp.dpf1UXaOt8 /tmp/tmp.6SkOZtwIv8 ++ return 0 + [[ 3 == \3 ]] + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-2.sql /tmp/tmp.1rIt3CjLpl/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-2.sql /tmp/tmp.1rIt3CjLpl/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-2.sql /tmp/tmp.1rIt3CjLpl/select-2.sql + desc 'test xtrabackup' + set +o xtrace ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.omZoCSDkHV ++ mktemp + local LAST_ERR=/tmp/tmp.0tTSLU8rro + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.omZoCSDkHV perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.0tTSLU8rro + rm /tmp/tmp.omZoCSDkHV /tmp/tmp.0tTSLU8rro + return 0 + patch_secret my-cluster-secrets xtrabackup dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=xtrabackup + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.ZyfxC5Ps0O ++ mktemp + local LAST_ERR=/tmp/tmp.wAVQFj0Ylm + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZyfxC5Ps0O secret/my-cluster-secrets patched + cat /tmp/tmp.wAVQFj0Ylm + rm /tmp/tmp.ZyfxC5Ps0O /tmp/tmp.wAVQFj0Ylm + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SvoqSnZ0Bp +++ mktemp ++ local LAST_ERR=/tmp/tmp.ixCtTHXOkg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SvoqSnZ0Bp ++ cat /tmp/tmp.ixCtTHXOkg ++ rm /tmp/tmp.SvoqSnZ0Bp /tmp/tmp.ixCtTHXOkg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PMAUvkn57c +++ mktemp ++ local LAST_ERR=/tmp/tmp.57YgQHcUmL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PMAUvkn57c ++ cat /tmp/tmp.57YgQHcUmL ++ rm /tmp/tmp.PMAUvkn57c /tmp/tmp.57YgQHcUmL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7jBk7sogFN +++ mktemp ++ local LAST_ERR=/tmp/tmp.jfY93ceWG8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7jBk7sogFN ++ cat /tmp/tmp.jfY93ceWG8 ++ rm /tmp/tmp.7jBk7sogFN /tmp/tmp.jfY93ceWG8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pjcDc6qAfQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.q8GnDegyoe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pjcDc6qAfQ ++ cat /tmp/tmp.q8GnDegyoe ++ rm /tmp/tmp.pjcDc6qAfQ /tmp/tmp.q8GnDegyoe ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FKyFapYF72 +++ mktemp ++ local LAST_ERR=/tmp/tmp.aSzLFRQYWF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FKyFapYF72 ++ cat /tmp/tmp.aSzLFRQYWF ++ rm /tmp/tmp.FKyFapYF72 /tmp/tmp.aSzLFRQYWF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iXExcnDiHR +++ mktemp ++ local LAST_ERR=/tmp/tmp.aSJHwcmaQV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iXExcnDiHR ++ cat /tmp/tmp.aSJHwcmaQV ++ rm /tmp/tmp.iXExcnDiHR /tmp/tmp.aSJHwcmaQV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eXsg0rQEAD +++ mktemp ++ local LAST_ERR=/tmp/tmp.2pZywY0w4O ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eXsg0rQEAD ++ cat /tmp/tmp.2pZywY0w4O ++ rm /tmp/tmp.eXsg0rQEAD /tmp/tmp.2pZywY0w4O ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dcTSHM9JGc +++ mktemp ++ local LAST_ERR=/tmp/tmp.UczeA0OJb0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dcTSHM9JGc ++ cat /tmp/tmp.UczeA0OJb0 ++ rm /tmp/tmp.dcTSHM9JGc /tmp/tmp.UczeA0OJb0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 7 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.40f3OWTpB9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.2nQ1qm66mC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.40f3OWTpB9 ++ cat /tmp/tmp.2nQ1qm66mC ++ rm /tmp/tmp.40f3OWTpB9 /tmp/tmp.2nQ1qm66mC ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oHvUUkuIDs +++ mktemp ++ local LAST_ERR=/tmp/tmp.XYBEMw0YGF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oHvUUkuIDs ++ cat /tmp/tmp.XYBEMw0YGF ++ rm /tmp/tmp.oHvUUkuIDs /tmp/tmp.XYBEMw0YGF ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.oKDIpZZOlC ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.IxdObjPxEu +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.oKDIpZZOlC +++++ cat /tmp/tmp.IxdObjPxEu +++++ rm /tmp/tmp.oKDIpZZOlC /tmp/tmp.IxdObjPxEu +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.1Q6ofAc8et ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.fQfsmS5jHS +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.1Q6ofAc8et +++++ cat /tmp/tmp.fQfsmS5jHS +++++ rm /tmp/tmp.1Q6ofAc8et /tmp/tmp.fQfsmS5jHS +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oMaLXFXYoD +++ mktemp ++ local LAST_ERR=/tmp/tmp.yLIlQkJbXy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oMaLXFXYoD ++ cat /tmp/tmp.yLIlQkJbXy ++ rm /tmp/tmp.oMaLXFXYoD /tmp/tmp.yLIlQkJbXy ++ return 0 + [[ 2 == \2 ]] + compare_mysql_cmd_local select-3 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 '' pxc + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 pxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-3.sql /tmp/tmp.1rIt3CjLpl/select-3.sql + desc 'test monitor' + set +o xtrace ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.FaoYnBk689 ++ mktemp + local LAST_ERR=/tmp/tmp.K6rJYfosyT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FaoYnBk689 secret/my-cluster-secrets patched + cat /tmp/tmp.K6rJYfosyT + rm /tmp/tmp.FaoYnBk689 /tmp/tmp.K6rJYfosyT + return 0 + wait_for_password_propagation my-cluster-secrets monitor + local secret=my-cluster-secrets + local user=monitor + local max_retry=240 ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.BYEvwkYo9H +++ mktemp ++ local LAST_ERR=/tmp/tmp.jNS8sFHopP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BYEvwkYo9H ++ cat /tmp/tmp.jNS8sFHopP ++ rm /tmp/tmp.BYEvwkYo9H /tmp/tmp.jNS8sFHopP ++ return 0 + local root_pass=test-password + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + echo 'Skipping dual password feature doesn'\''t work for 5.7. PXC 5.7 doesn'\''t support it!' Skipping dual password feature doesn't work for 5.7. PXC 5.7 doesn't support it! + return + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5COC4zetgN +++ mktemp ++ local LAST_ERR=/tmp/tmp.zthtiKp5st ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5COC4zetgN ++ cat /tmp/tmp.zthtiKp5st ++ rm /tmp/tmp.5COC4zetgN /tmp/tmp.zthtiKp5st ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6j5O8CC5Yv +++ mktemp ++ local LAST_ERR=/tmp/tmp.PlKhYs0YkA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6j5O8CC5Yv ++ cat /tmp/tmp.PlKhYs0YkA ++ rm /tmp/tmp.6j5O8CC5Yv /tmp/tmp.PlKhYs0YkA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.D3snUCgSoq +++ mktemp ++ local LAST_ERR=/tmp/tmp.NBTtqOqmf4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.D3snUCgSoq ++ cat /tmp/tmp.NBTtqOqmf4 ++ rm /tmp/tmp.D3snUCgSoq /tmp/tmp.NBTtqOqmf4 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0WtgrZ9wBR +++ mktemp ++ local LAST_ERR=/tmp/tmp.yw7WAl92M5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0WtgrZ9wBR ++ cat /tmp/tmp.yw7WAl92M5 ++ rm /tmp/tmp.0WtgrZ9wBR /tmp/tmp.yw7WAl92M5 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.VedX4IwM7f ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.cOelnzqKoH +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.VedX4IwM7f +++++ cat /tmp/tmp.cOelnzqKoH +++++ rm /tmp/tmp.VedX4IwM7f /tmp/tmp.cOelnzqKoH +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.05aXWui1dm ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.jRJtZJQ1b8 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.05aXWui1dm +++++ cat /tmp/tmp.jRJtZJQ1b8 +++++ rm /tmp/tmp.05aXWui1dm /tmp/tmp.jRJtZJQ1b8 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AMF9v4qC1X +++ mktemp ++ local LAST_ERR=/tmp/tmp.F6Lc0wjCWb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AMF9v4qC1X ++ cat /tmp/tmp.F6Lc0wjCWb ++ rm /tmp/tmp.AMF9v4qC1X /tmp/tmp.F6Lc0wjCWb ++ return 0 + [[ 2 == \2 ]] + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DoxFa0lyAP +++ mktemp ++ local LAST_ERR=/tmp/tmp.iiFU1vT7Ch ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DoxFa0lyAP ++ cat /tmp/tmp.iiFU1vT7Ch ++ rm /tmp/tmp.DoxFa0lyAP /tmp/tmp.iiFU1vT7Ch ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql /tmp/tmp.1rIt3CjLpl/select-4.sql + desc 'test operator' + set +o xtrace ----------------------------------------------------------------------------------- test operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets operator dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=operator + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.UpbfPJ2Tgf ++ mktemp + local LAST_ERR=/tmp/tmp.4sPlqNnNJ7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.UpbfPJ2Tgf secret/my-cluster-secrets patched + cat /tmp/tmp.4sPlqNnNJ7 + rm /tmp/tmp.UpbfPJ2Tgf /tmp/tmp.4sPlqNnNJ7 + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oh8sHP2y97 +++ mktemp ++ local LAST_ERR=/tmp/tmp.adV1UrHouQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oh8sHP2y97 ++ cat /tmp/tmp.adV1UrHouQ ++ rm /tmp/tmp.oh8sHP2y97 /tmp/tmp.adV1UrHouQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Az0gkfw9in +++ mktemp ++ local LAST_ERR=/tmp/tmp.FT29TTSd8M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Az0gkfw9in ++ cat /tmp/tmp.FT29TTSd8M ++ rm /tmp/tmp.Az0gkfw9in /tmp/tmp.FT29TTSd8M ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tQ80FCCIyH +++ mktemp ++ local LAST_ERR=/tmp/tmp.VryJxCW5hW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tQ80FCCIyH ++ cat /tmp/tmp.VryJxCW5hW ++ rm /tmp/tmp.tQ80FCCIyH /tmp/tmp.VryJxCW5hW ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.48xlpSC3aY ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.CLKGtnD3n6 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.48xlpSC3aY +++++ cat /tmp/tmp.CLKGtnD3n6 +++++ rm /tmp/tmp.48xlpSC3aY /tmp/tmp.CLKGtnD3n6 +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.FTVfl96uRN ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.bVS78gdB01 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.FTVfl96uRN +++++ cat /tmp/tmp.bVS78gdB01 +++++ rm /tmp/tmp.FTVfl96uRN /tmp/tmp.bVS78gdB01 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.akQYjxSYs9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.foc0SFvA6K ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.akQYjxSYs9 ++ cat /tmp/tmp.foc0SFvA6K ++ rm /tmp/tmp.akQYjxSYs9 /tmp/tmp.foc0SFvA6K ++ return 0 + [[ 2 == \2 ]] + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lG1m7lOCJ0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.P6n7aU4Ttc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lG1m7lOCJ0 ++ cat /tmp/tmp.P6n7aU4Ttc ++ rm /tmp/tmp.lG1m7lOCJ0 /tmp/tmp.P6n7aU4Ttc ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql /tmp/tmp.1rIt3CjLpl/select-4.sql + desc 'change secret name' + set +o xtrace ----------------------------------------------------------------------------------- change secret name ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.vOH3bo1iOw ++ mktemp + local LAST_ERR=/tmp/tmp.WxFejibj39 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vOH3bo1iOw perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.WxFejibj39 + rm /tmp/tmp.vOH3bo1iOw /tmp/tmp.WxFejibj39 + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xq8AJJRCS9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZBxTnwaavx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xq8AJJRCS9 ++ cat /tmp/tmp.ZBxTnwaavx ++ rm /tmp/tmp.xq8AJJRCS9 /tmp/tmp.ZBxTnwaavx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nPoHZlPsuT +++ mktemp ++ local LAST_ERR=/tmp/tmp.0UcLsGeIsg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nPoHZlPsuT ++ cat /tmp/tmp.0UcLsGeIsg ++ rm /tmp/tmp.nPoHZlPsuT /tmp/tmp.0UcLsGeIsg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HYxZw9kEhf +++ mktemp ++ local LAST_ERR=/tmp/tmp.qzrElVHpji ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HYxZw9kEhf ++ cat /tmp/tmp.qzrElVHpji ++ rm /tmp/tmp.HYxZw9kEhf /tmp/tmp.qzrElVHpji ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XtNkQaBJrf +++ mktemp ++ local LAST_ERR=/tmp/tmp.P59tlHm0O9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XtNkQaBJrf ++ cat /tmp/tmp.P59tlHm0O9 ++ rm /tmp/tmp.XtNkQaBJrf /tmp/tmp.P59tlHm0O9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0cEVMzFVOb +++ mktemp ++ local LAST_ERR=/tmp/tmp.MTwj5bsWj0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0cEVMzFVOb ++ cat /tmp/tmp.MTwj5bsWj0 ++ rm /tmp/tmp.0cEVMzFVOb /tmp/tmp.MTwj5bsWj0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QNwTntD3KN +++ mktemp ++ local LAST_ERR=/tmp/tmp.Os4jbPMvCP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QNwTntD3KN ++ cat /tmp/tmp.Os4jbPMvCP ++ rm /tmp/tmp.QNwTntD3KN /tmp/tmp.Os4jbPMvCP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5IUXS38Y4B +++ mktemp ++ local LAST_ERR=/tmp/tmp.1gDvrOwH7M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5IUXS38Y4B ++ cat /tmp/tmp.1gDvrOwH7M ++ rm /tmp/tmp.5IUXS38Y4B /tmp/tmp.1gDvrOwH7M ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ddNRIFcsij +++ mktemp ++ local LAST_ERR=/tmp/tmp.XoCDbRSfTj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ddNRIFcsij ++ cat /tmp/tmp.XoCDbRSfTj ++ rm /tmp/tmp.ddNRIFcsij /tmp/tmp.XoCDbRSfTj ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.YqPEGiB0HU ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.vE4kORLqeY +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.YqPEGiB0HU +++++ cat /tmp/tmp.vE4kORLqeY +++++ rm /tmp/tmp.YqPEGiB0HU /tmp/tmp.vE4kORLqeY +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.OTNhSokmOE ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.jqc7OKrsyi +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.OTNhSokmOE +++++ cat /tmp/tmp.jqc7OKrsyi +++++ rm /tmp/tmp.OTNhSokmOE /tmp/tmp.jqc7OKrsyi +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GxzVzTS0ry +++ mktemp ++ local LAST_ERR=/tmp/tmp.zCpXdlyVoj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GxzVzTS0ry ++ cat /tmp/tmp.zCpXdlyVoj ++ rm /tmp/tmp.GxzVzTS0ry /tmp/tmp.zCpXdlyVoj ++ return 0 + [[ 2 == \2 ]] + desc 'test new operator' + set +o xtrace ----------------------------------------------------------------------------------- test new operator ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + patch_secret my-cluster-secrets-2 operator dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets-2 + local key=operator + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.UQuG4bbl96 ++ mktemp + local LAST_ERR=/tmp/tmp.ILcVNHRSOO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.UQuG4bbl96 secret/my-cluster-secrets-2 patched + cat /tmp/tmp.ILcVNHRSOO + rm /tmp/tmp.UQuG4bbl96 /tmp/tmp.ILcVNHRSOO + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2pFcvgTm8C +++ mktemp ++ local LAST_ERR=/tmp/tmp.sjM06BAEns ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2pFcvgTm8C ++ cat /tmp/tmp.sjM06BAEns ++ rm /tmp/tmp.2pFcvgTm8C /tmp/tmp.sjM06BAEns ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uSMES8SVk0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.xo4tg2BwSi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uSMES8SVk0 ++ cat /tmp/tmp.xo4tg2BwSi ++ rm /tmp/tmp.uSMES8SVk0 /tmp/tmp.xo4tg2BwSi ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.p67i0SYb3j ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.axVp28Xwul +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.p67i0SYb3j +++++ cat /tmp/tmp.axVp28Xwul +++++ rm /tmp/tmp.p67i0SYb3j /tmp/tmp.axVp28Xwul +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.k2OooZlvby ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.ALpeUjdUoi +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.k2OooZlvby +++++ cat /tmp/tmp.ALpeUjdUoi +++++ rm /tmp/tmp.k2OooZlvby /tmp/tmp.ALpeUjdUoi +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bMiT651kgx +++ mktemp ++ local LAST_ERR=/tmp/tmp.ehaAvlIpj7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bMiT651kgx ++ cat /tmp/tmp.ehaAvlIpj7 ++ rm /tmp/tmp.bMiT651kgx /tmp/tmp.ehaAvlIpj7 ++ return 0 + [[ 2 == \2 ]] + sleep 20 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B7wZIeG9Cp +++ mktemp ++ local LAST_ERR=/tmp/tmp.WxDLAMyQa4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B7wZIeG9Cp ++ cat /tmp/tmp.WxDLAMyQa4 ++ rm /tmp/tmp.B7wZIeG9Cp /tmp/tmp.WxDLAMyQa4 ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql /tmp/tmp.1rIt3CjLpl/select-4.sql ++ getSecretData my-cluster-secrets-2 root ++ local secretName=my-cluster-secrets-2 ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.SZiNLNtoMa +++ mktemp ++ local LAST_ERR=/tmp/tmp.qVbSCES8z0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SZiNLNtoMa ++ cat /tmp/tmp.qVbSCES8z0 ++ rm /tmp/tmp.SZiNLNtoMa /tmp/tmp.qVbSCES8z0 ++ return 0 + newpass='3k3%jbdOuUm6xeV[' + desc 'test new users sync' + set +o xtrace ----------------------------------------------------------------------------------- test new users sync ----------------------------------------------------------------------------------- + run_mysql 'CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''3k3%jbdOuUm6xeV['\'';' '-h some-name-pxc -uroot -p'\''3k3%jbdOuUm6xeV['\''' + local 'command=CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''3k3%jbdOuUm6xeV['\'';' + local 'uri=-h some-name-pxc -uroot -p'\''3k3%jbdOuUm6xeV['\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ln2MEozHZv +++ mktemp ++ local LAST_ERR=/tmp/tmp.GgUiVKQCfn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ln2MEozHZv ++ cat /tmp/tmp.GgUiVKQCfn ++ rm /tmp/tmp.Ln2MEozHZv /tmp/tmp.GgUiVKQCfn ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + sleep 40 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''3k3%jbdOuUm6xeV['\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''3k3%jbdOuUm6xeV['\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''3k3%jbdOuUm6xeV['\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''3k3%jbdOuUm6xeV['\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1cIsOeAv3R +++ mktemp ++ local LAST_ERR=/tmp/tmp.ycwYZMjn9l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1cIsOeAv3R ++ cat /tmp/tmp.ycwYZMjn9l ++ rm /tmp/tmp.1cIsOeAv3R /tmp/tmp.ycwYZMjn9l ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql /tmp/tmp.1rIt3CjLpl/select-4.sql ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.o1gLONvu8R +++ mktemp ++ local LAST_ERR=/tmp/tmp.K22d9XPC1N ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o1gLONvu8R ++ cat /tmp/tmp.K22d9XPC1N ++ rm /tmp/tmp.o1gLONvu8R /tmp/tmp.K22d9XPC1N ++ return 0 + pass=test-password2 + desc 'check secret without operator' + set +o xtrace ----------------------------------------------------------------------------------- check secret without operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.qlM1yulwfn ++ mktemp + local LAST_ERR=/tmp/tmp.jwY2GSFA5S + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qlM1yulwfn secret/my-cluster-secrets-2 configured + cat /tmp/tmp.jwY2GSFA5S Warning: resource secrets/my-cluster-secrets-2 is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.qlM1yulwfn /tmp/tmp.jwY2GSFA5S + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XtWP4HN9Yo +++ mktemp ++ local LAST_ERR=/tmp/tmp.dpyBKI7WgI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XtWP4HN9Yo ++ cat /tmp/tmp.dpyBKI7WgI ++ rm /tmp/tmp.XtWP4HN9Yo /tmp/tmp.dpyBKI7WgI ++ return 0 + client_pod=pxc-client-7b7f8bcff9-mllsx + wait_pod pxc-client-7b7f8bcff9-mllsx + local pod=pxc-client-7b7f8bcff9-mllsx + local max_retry=480 + local ns= ++ echo pxc-client-7b7f8bcff9-mllsx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b7f8bcff9-mllsx condition met pxc-client-7b7f8bcff9-mllsx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.1rIt3CjLpl/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql /tmp/tmp.1rIt3CjLpl/select-4.sql --- /mnt/jenkins/workspace/cloud-pxc-operator_PR-1781/e2e-tests/users/compare/select-4.sql 2024-08-08 01:56:03.895328536 +0000 +++ /tmp/tmp.1rIt3CjLpl/select-4.sql 2024-08-08 04:36:51.949959944 +0000 @@ -1,76 +1,2 @@ -CHARACTER_SETS -CLIENT_STATISTICS -COLLATIONS -COLLATION_CHARACTER_SET_APPLICABILITY -COLUMNS -COLUMN_PRIVILEGES -INDEX_STATISTICS -ENGINES -EVENTS -FILES -GLOBAL_STATUS -GLOBAL_TEMPORARY_TABLES -GLOBAL_VARIABLES -KEY_COLUMN_USAGE -OPTIMIZER_TRACE -PARAMETERS -PARTITIONS -PLUGINS -PROCESSLIST -PROFILING -REFERENTIAL_CONSTRAINTS -ROUTINES -SCHEMATA -SCHEMA_PRIVILEGES -SESSION_STATUS -SESSION_VARIABLES -STATISTICS -TABLES -TABLESPACES -TABLE_CONSTRAINTS -TABLE_PRIVILEGES -TABLE_STATISTICS -TEMPORARY_TABLES -THREAD_STATISTICS -TRIGGERS -USER_PRIVILEGES -USER_STATISTICS -VIEWS -INNODB_CMP -INNODB_CMP_RESET -INNODB_SYS_DATAFILES -XTRADB_READ_VIEW -INNODB_SYS_TABLESTATS -XTRADB_RSEG -INNODB_BUFFER_PAGE -INNODB_TRX -INNODB_CMP_PER_INDEX -INNODB_METRICS -INNODB_FT_DELETED -INNODB_LOCKS -INNODB_LOCK_WAITS -XTRADB_INTERNAL_HASH_TABLES -INNODB_TABLESPACES_ENCRYPTION -INNODB_CMPMEM_RESET -INNODB_SYS_FIELDS -XTRADB_ZIP_DICT -INNODB_TABLESPACES_SCRUBBING -INNODB_TEMP_TABLE_INFO -INNODB_FT_INDEX_TABLE -INNODB_CMPMEM -INNODB_SYS_TABLESPACES -INNODB_CMP_PER_INDEX_RESET -INNODB_SYS_FOREIGN_COLS -INNODB_FT_INDEX_CACHE -INNODB_BUFFER_POOL_STATS -INNODB_FT_BEING_DELETED -INNODB_SYS_FOREIGN -INNODB_BUFFER_PAGE_LRU -INNODB_FT_DEFAULT_STOPWORD -INNODB_SYS_TABLES -INNODB_SYS_COLUMNS -INNODB_FT_CONFIG -XTRADB_ZIP_DICT_COLS -INNODB_SYS_INDEXES -INNODB_SYS_VIRTUAL -INNODB_CHANGED_PAGES +ERROR 2005 (HY000): Unknown MySQL server host 'some-name-proxysql' (2) +command terminated with exit code 1