Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/logs/users-5-7.log WARNING: version difference between client (1.29) and server (1.25) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.29) and server (1.25) exceeds the supported minor version skew of +/-1 + create_infra users-14045 + local ns=users-14045 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n users-976 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.W4UROMJ5I9 ++ mktemp + local LAST_ERR=/tmp/tmp.142pCe7t54 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.W4UROMJ5I9 perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.142pCe7t54 + rm /tmp/tmp.W4UROMJ5I9 /tmp/tmp.142pCe7t54 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.9LplPUEkKf ++ mktemp + local LAST_ERR=/tmp/tmp.Q86RdJkdyo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9LplPUEkKf No resources found + cat /tmp/tmp.Q86RdJkdyo + rm /tmp/tmp.9LplPUEkKf /tmp/tmp.Q86RdJkdyo + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.XbKDXf5brE ++ mktemp + local LAST_ERR=/tmp/tmp.UD956Ndiyb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.XbKDXf5brE No resources found + cat /tmp/tmp.UD956Ndiyb + rm /tmp/tmp.XbKDXf5brE /tmp/tmp.UD956Ndiyb + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' + '[' -n '' ']' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + xargs kubectl delete ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.xYNcMcD8N0 + local LAST_OUT=/tmp/tmp.X6gl1NC45C ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.mBfFSKg1m6 + local exit_status=0 + local LAST_ERR=/tmp/tmp.Pv3RpZ7SdC + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xYNcMcD8N0 + cat /tmp/tmp.mBfFSKg1m6 + rm /tmp/tmp.xYNcMcD8N0 /tmp/tmp.mBfFSKg1m6 + return 0 namespace "users-976" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.X6gl1NC45C namespace "pxc-operator" deleted + cat /tmp/tmp.Pv3RpZ7SdC + rm /tmp/tmp.X6gl1NC45C /tmp/tmp.Pv3RpZ7SdC + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'namespace/pxc-operator - ' namespace/pxc-operator - + set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.LKB7Rl5k0w ++ mktemp + local LAST_ERR=/tmp/tmp.XkKH4mSTv5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LKB7Rl5k0w namespace/pxc-operator created + cat /tmp/tmp.XkKH4mSTv5 + rm /tmp/tmp.LKB7Rl5k0w /tmp/tmp.XkKH4mSTv5 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Qhl2bnPiq7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yDn90rHqpc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Qhl2bnPiq7 ++ cat /tmp/tmp.yDn90rHqpc ++ rm /tmp/tmp.Qhl2bnPiq7 /tmp/tmp.yDn90rHqpc ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1421-86089695-4-cluster1 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.adO1PFtpbL ++ mktemp + local LAST_ERR=/tmp/tmp.H4uz6Tl4Zc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1421-86089695-4-cluster1 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.adO1PFtpbL Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1421-86089695-4-cluster1" modified. + cat /tmp/tmp.H4uz6Tl4Zc + rm /tmp/tmp.adO1PFtpbL /tmp/tmp.H4uz6Tl4Zc + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.VRNg9B4KNf ++ mktemp + local LAST_ERR=/tmp/tmp.I8ckedyKuI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VRNg9B4KNf customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.I8ckedyKuI + rm /tmp/tmp.VRNg9B4KNf /tmp/tmp.I8ckedyKuI + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.qrSRnBlpzT ++ mktemp + local LAST_ERR=/tmp/tmp.8Rw1mIQ515 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qrSRnBlpzT clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.8Rw1mIQ515 + rm /tmp/tmp.qrSRnBlpzT /tmp/tmp.8Rw1mIQ515 + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/deploy/cw-operator.yaml + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1421-86089695^' + kubectl_bin apply -f - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "DEBUG"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - ++ mktemp + local LAST_OUT=/tmp/tmp.AXqi1JVbLv ++ mktemp + local LAST_ERR=/tmp/tmp.bNiZRXbhqQ + local exit_status=0 + sed -e 's^failureThreshold: .*^failureThreshold: 10^' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.AXqi1JVbLv deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.bNiZRXbhqQ + rm /tmp/tmp.AXqi1JVbLv /tmp/tmp.bNiZRXbhqQ + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.X3mkXxlfVM ++ mktemp + local LAST_ERR=/tmp/tmp.WPxLW19EGQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.X3mkXxlfVM pod/percona-xtradb-cluster-operator-5c699b45b5-7nst4 condition met + cat /tmp/tmp.WPxLW19EGQ + rm /tmp/tmp.X3mkXxlfVM /tmp/tmp.WPxLW19EGQ + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.OJ054rBVef +++ mktemp ++ local LAST_ERR=/tmp/tmp.acUT1zbMPV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OJ054rBVef ++ cat /tmp/tmp.acUT1zbMPV ++ rm /tmp/tmp.OJ054rBVef /tmp/tmp.acUT1zbMPV ++ return 0 + wait_pod percona-xtradb-cluster-operator-5c699b45b5-7nst4 480 pxc-operator + local pod=percona-xtradb-cluster-operator-5c699b45b5-7nst4 + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-5c699b45b5-7nst4 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-5c699b45b5-7nst4 condition met percona-xtradb-cluster-operator-5c699b45b5-7nst4.Ok + sleep 3 + create_namespace users-14045 + local namespace=users-14045 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' + '[' -n '' ']' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces users-14045' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces users-14045 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace users-14045 + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' ++ mktemp + local LAST_OUT=/tmp/tmp.lMFzOqrUGN ++ mktemp + local LAST_OUT=/tmp/tmp.J9kV4EyHp5 ++ mktemp + local LAST_ERR=/tmp/tmp.1InKMFRiBk + local exit_status=0 + local LAST_ERR=/tmp/tmp.yhP3nyrQRO + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-14045 + for i in '$(seq 0 2)' + set +e + kubectl get ns + awk '{print$1}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.lMFzOqrUGN + cat /tmp/tmp.1InKMFRiBk + rm /tmp/tmp.lMFzOqrUGN /tmp/tmp.1InKMFRiBk + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-14045 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-14045 Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.J9kV4EyHp5 + cat /tmp/tmp.yhP3nyrQRO Error from server (NotFound): namespaces "users-14045" not found + rm /tmp/tmp.J9kV4EyHp5 /tmp/tmp.yhP3nyrQRO + return 1 + : + wait_for_delete namespace/users-14045 + local res=namespace/users-14045 + echo -n 'namespace/users-14045 - ' namespace/users-14045 - + set +o xtrace Error from server (NotFound): namespaces "users-14045" not found + desc 'create namespace users-14045' + set +o xtrace ----------------------------------------------------------------------------------- create namespace users-14045 ----------------------------------------------------------------------------------- + kubectl_bin create namespace users-14045 ++ mktemp + local LAST_OUT=/tmp/tmp.L52faOA7O6 ++ mktemp + local LAST_ERR=/tmp/tmp.zPFdXsnkSZ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace users-14045 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.L52faOA7O6 namespace/users-14045 created + cat /tmp/tmp.zPFdXsnkSZ + rm /tmp/tmp.L52faOA7O6 /tmp/tmp.zPFdXsnkSZ + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.TvBHWRFbsP +++ mktemp ++ local LAST_ERR=/tmp/tmp.MbSCayFAwn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TvBHWRFbsP ++ cat /tmp/tmp.MbSCayFAwn ++ rm /tmp/tmp.TvBHWRFbsP /tmp/tmp.MbSCayFAwn ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1421-86089695-4-cluster1 --namespace=users-14045 ++ mktemp + local LAST_OUT=/tmp/tmp.hIu0UOOBRO ++ mktemp + local LAST_ERR=/tmp/tmp.Pcv2tnRXkl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1421-86089695-4-cluster1 --namespace=users-14045 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hIu0UOOBRO Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1421-86089695-4-cluster1" modified. + cat /tmp/tmp.Pcv2tnRXkl + rm /tmp/tmp.hIu0UOOBRO /tmp/tmp.Pcv2tnRXkl + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.pFU5i27xu5 ++ mktemp + local LAST_ERR=/tmp/tmp.CtfaJxgAA0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pFU5i27xu5 secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.CtfaJxgAA0 + rm /tmp/tmp.pFU5i27xu5 /tmp/tmp.CtfaJxgAA0 + return 0 + desc 'create PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- + newpass=test-password ++ echo -n test-password ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZA== + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.dQsQm21UQA ++ mktemp + local LAST_ERR=/tmp/tmp.HszGDjoltQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dQsQm21UQA secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.HszGDjoltQ + rm /tmp/tmp.dQsQm21UQA /tmp/tmp.HszGDjoltQ + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/client.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/client.yml ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + local LAST_OUT=/tmp/tmp.W4XzCeIp1R + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1421-86089695#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ mktemp + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_ERR=/tmp/tmp.JzQiYfUAeF + local exit_status=0 + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-14045~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.W4XzCeIp1R deployment.apps/pxc-client created + cat /tmp/tmp.JzQiYfUAeF + rm /tmp/tmp.W4XzCeIp1R /tmp/tmp.JzQiYfUAeF + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p\-\c\l\o\u\d ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/some-name.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1421-86089695#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + local LAST_OUT=/tmp/tmp.bzkydLVBJg + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/conf/some-name.yml + local LAST_ERR=/tmp/tmp.G9QOGxz2zl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-14045~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.bzkydLVBJg perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.G9QOGxz2zl + rm /tmp/tmp.bzkydLVBJg /tmp/tmp.G9QOGxz2zl + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.w1UJY8WvcH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.aG4MXoXRZb +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.w1UJY8WvcH +++ cat /tmp/tmp.aG4MXoXRZb +++ rm /tmp/tmp.w1UJY8WvcH /tmp/tmp.aG4MXoXRZb +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.E8anwGHhNz ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vRZVlmx4JK +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.E8anwGHhNz +++ cat /tmp/tmp.vRZVlmx4JK +++ rm /tmp/tmp.E8anwGHhNz /tmp/tmp.vRZVlmx4JK +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-14045 ++ mktemp + local LAST_OUT=/tmp/tmp.IJiywQEKkW ++ mktemp + local LAST_ERR=/tmp/tmp.7tbBdIMa5z + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-14045 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-14045 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-14045 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.IJiywQEKkW + cat /tmp/tmp.7tbBdIMa5z error: no matching resources found + rm /tmp/tmp.IJiywQEKkW /tmp/tmp.7tbBdIMa5z + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met some-name-proxysql-0.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met some-name-pxc-0.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met some-name-pxc-1.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met some-name-pxc-2.Ok + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ICl0kMq5RP +++ mktemp ++ local LAST_ERR=/tmp/tmp.3TqsmDFJep ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ICl0kMq5RP ++ cat /tmp/tmp.3TqsmDFJep Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.ICl0kMq5RP /tmp/tmp.3TqsmDFJep ++ return 0 + [[ -n '' ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -proot_password -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TJHXtwIAFn +++ mktemp ++ local LAST_ERR=/tmp/tmp.2HEvPRwGIc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TJHXtwIAFn ++ cat /tmp/tmp.2HEvPRwGIc ++ rm /tmp/tmp.TJHXtwIAFn /tmp/tmp.2HEvPRwGIc ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -proot_password -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3WNqBraUKB +++ mktemp ++ local LAST_ERR=/tmp/tmp.KewWxVMVhL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3WNqBraUKB ++ cat /tmp/tmp.KewWxVMVhL ++ rm /tmp/tmp.3WNqBraUKB /tmp/tmp.KewWxVMVhL ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KjL6vfz6Va +++ mktemp ++ local LAST_ERR=/tmp/tmp.EUOj5KaGGY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KjL6vfz6Va ++ cat /tmp/tmp.EUOj5KaGGY ++ rm /tmp/tmp.KjL6vfz6Va /tmp/tmp.EUOj5KaGGY ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-1.sql /tmp/tmp.xOj8EZf0bc/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bCGyJLjkdP +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ag8sO8bhmt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bCGyJLjkdP ++ cat /tmp/tmp.Ag8sO8bhmt ++ rm /tmp/tmp.bCGyJLjkdP /tmp/tmp.Ag8sO8bhmt ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-1.sql /tmp/tmp.xOj8EZf0bc/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.X4gVk3VmOa +++ mktemp ++ local LAST_ERR=/tmp/tmp.6R6YXoez3S ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.X4gVk3VmOa ++ cat /tmp/tmp.6R6YXoez3S ++ rm /tmp/tmp.X4gVk3VmOa /tmp/tmp.6R6YXoez3S ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-1.sql /tmp/tmp.xOj8EZf0bc/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NOcReUM8ci +++ mktemp ++ local LAST_ERR=/tmp/tmp.wnZFaGJxBQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NOcReUM8ci ++ cat /tmp/tmp.wnZFaGJxBQ Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.NOcReUM8ci /tmp/tmp.wnZFaGJxBQ ++ return 0 + '[' '' ']' + desc 'test root' + set +o xtrace ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets root dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=root + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.RQyiRN7QnN ++ mktemp + local LAST_ERR=/tmp/tmp.8e25OOwVXR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RQyiRN7QnN secret/my-cluster-secrets patched + cat /tmp/tmp.8e25OOwVXR + rm /tmp/tmp.RQyiRN7QnN /tmp/tmp.8e25OOwVXR + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o6wlaH7Jtr +++ mktemp ++ local LAST_ERR=/tmp/tmp.pfQ2VqadHh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o6wlaH7Jtr ++ cat /tmp/tmp.pfQ2VqadHh ++ rm /tmp/tmp.o6wlaH7Jtr /tmp/tmp.pfQ2VqadHh ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql /tmp/tmp.xOj8EZf0bc/select-4.sql + desc 'test proxyadmin' + set +o xtrace ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.d4imSc9hv2 ++ mktemp + local LAST_ERR=/tmp/tmp.wj5myVD7of + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.d4imSc9hv2 perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.wj5myVD7of + rm /tmp/tmp.d4imSc9hv2 /tmp/tmp.wj5myVD7of + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dzMy65oakB +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lgi0tvDq8W ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dzMy65oakB ++ cat /tmp/tmp.Lgi0tvDq8W ++ rm /tmp/tmp.dzMy65oakB /tmp/tmp.Lgi0tvDq8W ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5H3MOCBNWe +++ mktemp ++ local LAST_ERR=/tmp/tmp.4QIUvj7d3i ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5H3MOCBNWe ++ cat /tmp/tmp.4QIUvj7d3i ++ rm /tmp/tmp.5H3MOCBNWe /tmp/tmp.4QIUvj7d3i ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.eK5xZeidoZ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.ZcJjxt3Ogy +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.eK5xZeidoZ +++++ cat /tmp/tmp.ZcJjxt3Ogy +++++ rm /tmp/tmp.eK5xZeidoZ /tmp/tmp.ZcJjxt3Ogy +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.nYqgrunJkK ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.My8L7Drcup +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.nYqgrunJkK +++++ cat /tmp/tmp.My8L7Drcup +++++ rm /tmp/tmp.nYqgrunJkK /tmp/tmp.My8L7Drcup +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f2jTqUPD4u +++ mktemp ++ local LAST_ERR=/tmp/tmp.1nIRSGAHo7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f2jTqUPD4u ++ cat /tmp/tmp.1nIRSGAHo7 ++ rm /tmp/tmp.f2jTqUPD4u /tmp/tmp.1nIRSGAHo7 ++ return 0 + [[ 3 == \3 ]] + patch_secret my-cluster-secrets proxyadmin dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=proxyadmin + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.mEz5DORa4L ++ mktemp + local LAST_ERR=/tmp/tmp.5bQqRj79ch + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mEz5DORa4L secret/my-cluster-secrets patched + cat /tmp/tmp.5bQqRj79ch + rm /tmp/tmp.mEz5DORa4L /tmp/tmp.5bQqRj79ch + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4XEPRlgxsx +++ mktemp ++ local LAST_ERR=/tmp/tmp.2iYnOBwtDm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4XEPRlgxsx ++ cat /tmp/tmp.2iYnOBwtDm ++ rm /tmp/tmp.4XEPRlgxsx /tmp/tmp.2iYnOBwtDm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rTskzkibXK +++ mktemp ++ local LAST_ERR=/tmp/tmp.Bab7cMdWau ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rTskzkibXK ++ cat /tmp/tmp.Bab7cMdWau ++ rm /tmp/tmp.rTskzkibXK /tmp/tmp.Bab7cMdWau ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3pyW0Uv6az +++ mktemp ++ local LAST_ERR=/tmp/tmp.3ILl8xsv6c ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3pyW0Uv6az ++ cat /tmp/tmp.3ILl8xsv6c ++ rm /tmp/tmp.3pyW0Uv6az /tmp/tmp.3ILl8xsv6c ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.pqY5nNJMQi ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.ZFREvFHSaf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.pqY5nNJMQi +++++ cat /tmp/tmp.ZFREvFHSaf +++++ rm /tmp/tmp.pqY5nNJMQi /tmp/tmp.ZFREvFHSaf +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.ysCWGTOgWl ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.5cuP1trkAz +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.ysCWGTOgWl +++++ cat /tmp/tmp.5cuP1trkAz +++++ rm /tmp/tmp.ysCWGTOgWl /tmp/tmp.5cuP1trkAz +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zxngllok7E +++ mktemp ++ local LAST_ERR=/tmp/tmp.74MccxAFrs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Zxngllok7E ++ cat /tmp/tmp.74MccxAFrs ++ rm /tmp/tmp.Zxngllok7E /tmp/tmp.74MccxAFrs ++ return 0 + [[ 3 == \3 ]] + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-2.sql /tmp/tmp.xOj8EZf0bc/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-2.sql /tmp/tmp.xOj8EZf0bc/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-2.sql /tmp/tmp.xOj8EZf0bc/select-2.sql + desc 'test xtrabackup' + set +o xtrace ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.pRP9DpV8Bt ++ mktemp + local LAST_ERR=/tmp/tmp.ZcuVTrKzPu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pRP9DpV8Bt perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.ZcuVTrKzPu + rm /tmp/tmp.pRP9DpV8Bt /tmp/tmp.ZcuVTrKzPu + return 0 + patch_secret my-cluster-secrets xtrabackup dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=xtrabackup + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.bfKA1StK0q ++ mktemp + local LAST_ERR=/tmp/tmp.MT6QXHjpXT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.bfKA1StK0q secret/my-cluster-secrets patched + cat /tmp/tmp.MT6QXHjpXT + rm /tmp/tmp.bfKA1StK0q /tmp/tmp.MT6QXHjpXT + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qaRFoJ3ps7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jeH1bXbWIh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qaRFoJ3ps7 ++ cat /tmp/tmp.jeH1bXbWIh ++ rm /tmp/tmp.qaRFoJ3ps7 /tmp/tmp.jeH1bXbWIh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PurocKXYw5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.c61KAKrktF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PurocKXYw5 ++ cat /tmp/tmp.c61KAKrktF ++ rm /tmp/tmp.PurocKXYw5 /tmp/tmp.c61KAKrktF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3VP39ODFN7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3IeVYBUGKc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3VP39ODFN7 ++ cat /tmp/tmp.3IeVYBUGKc ++ rm /tmp/tmp.3VP39ODFN7 /tmp/tmp.3IeVYBUGKc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R6bOyCN12l +++ mktemp ++ local LAST_ERR=/tmp/tmp.uSUNIfhcCP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.R6bOyCN12l ++ cat /tmp/tmp.uSUNIfhcCP ++ rm /tmp/tmp.R6bOyCN12l /tmp/tmp.uSUNIfhcCP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vFfeHT9egI +++ mktemp ++ local LAST_ERR=/tmp/tmp.3dQrvdtJjM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vFfeHT9egI ++ cat /tmp/tmp.3dQrvdtJjM ++ rm /tmp/tmp.vFfeHT9egI /tmp/tmp.3dQrvdtJjM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Br0a4uSKs3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.8Rmz2NXfZp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Br0a4uSKs3 ++ cat /tmp/tmp.8Rmz2NXfZp ++ rm /tmp/tmp.Br0a4uSKs3 /tmp/tmp.8Rmz2NXfZp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gHShhAeUOy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jq32uAyF8m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gHShhAeUOy ++ cat /tmp/tmp.Jq32uAyF8m ++ rm /tmp/tmp.gHShhAeUOy /tmp/tmp.Jq32uAyF8m ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PyOyzq0IN1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ItSowhSSnt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PyOyzq0IN1 ++ cat /tmp/tmp.ItSowhSSnt ++ rm /tmp/tmp.PyOyzq0IN1 /tmp/tmp.ItSowhSSnt ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uZmDyn0hVS +++ mktemp ++ local LAST_ERR=/tmp/tmp.8tuSzcNMhh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uZmDyn0hVS ++ cat /tmp/tmp.8tuSzcNMhh ++ rm /tmp/tmp.uZmDyn0hVS /tmp/tmp.8tuSzcNMhh ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.bGfE8DlxNi ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.lA6ZszKZJq +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.bGfE8DlxNi +++++ cat /tmp/tmp.lA6ZszKZJq +++++ rm /tmp/tmp.bGfE8DlxNi /tmp/tmp.lA6ZszKZJq +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.WdFtyDSul9 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.IjqIWGksZv +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.WdFtyDSul9 +++++ cat /tmp/tmp.IjqIWGksZv +++++ rm /tmp/tmp.WdFtyDSul9 /tmp/tmp.IjqIWGksZv +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.isqwcUylt1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cONqK9rllZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.isqwcUylt1 ++ cat /tmp/tmp.cONqK9rllZ ++ rm /tmp/tmp.isqwcUylt1 /tmp/tmp.cONqK9rllZ ++ return 0 + [[ 2 == \2 ]] + compare_mysql_cmd_local select-3 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 '' pxc + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 pxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-3.sql /tmp/tmp.xOj8EZf0bc/select-3.sql + desc 'test monitor' + set +o xtrace ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.yk20Af0dms ++ mktemp + local LAST_ERR=/tmp/tmp.wSoTvKkz8s + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yk20Af0dms secret/my-cluster-secrets patched + cat /tmp/tmp.wSoTvKkz8s + rm /tmp/tmp.yk20Af0dms /tmp/tmp.wSoTvKkz8s + return 0 + wait_for_password_propagation my-cluster-secrets monitor + local secret=my-cluster-secrets + local user=monitor + local max_retry=240 ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.pH1MgaUBoP +++ mktemp ++ local LAST_ERR=/tmp/tmp.VZ7eoO8U5y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pH1MgaUBoP ++ cat /tmp/tmp.VZ7eoO8U5y ++ rm /tmp/tmp.pH1MgaUBoP /tmp/tmp.VZ7eoO8U5y ++ return 0 + local root_pass=test-password + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + echo 'Skipping dual password feature doesn'\''t work for 5.7. PXC 5.7 doesn'\''t support it!' Skipping dual password feature doesn't work for 5.7. PXC 5.7 doesn't support it! + return + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Qy3cnuETSE +++ mktemp ++ local LAST_ERR=/tmp/tmp.OXh7DmQZgu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Qy3cnuETSE ++ cat /tmp/tmp.OXh7DmQZgu ++ rm /tmp/tmp.Qy3cnuETSE /tmp/tmp.OXh7DmQZgu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mv75K45B9a +++ mktemp ++ local LAST_ERR=/tmp/tmp.vxybofWJHv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mv75K45B9a ++ cat /tmp/tmp.vxybofWJHv ++ rm /tmp/tmp.mv75K45B9a /tmp/tmp.vxybofWJHv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vnZOw6JveQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.AYWmNBjoCl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vnZOw6JveQ ++ cat /tmp/tmp.AYWmNBjoCl ++ rm /tmp/tmp.vnZOw6JveQ /tmp/tmp.AYWmNBjoCl ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tH9X5ilhdQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.QxVHvIej3M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tH9X5ilhdQ ++ cat /tmp/tmp.QxVHvIej3M ++ rm /tmp/tmp.tH9X5ilhdQ /tmp/tmp.QxVHvIej3M ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.cFS4Q2StyN ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.ZBw2mJjSQD +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.cFS4Q2StyN +++++ cat /tmp/tmp.ZBw2mJjSQD +++++ rm /tmp/tmp.cFS4Q2StyN /tmp/tmp.ZBw2mJjSQD +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Fw2qimWjvc ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.KiwvfvgzRF +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Fw2qimWjvc +++++ cat /tmp/tmp.KiwvfvgzRF +++++ rm /tmp/tmp.Fw2qimWjvc /tmp/tmp.KiwvfvgzRF +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gLsjSmW6BS +++ mktemp ++ local LAST_ERR=/tmp/tmp.VvVyrRj3q7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gLsjSmW6BS ++ cat /tmp/tmp.VvVyrRj3q7 ++ rm /tmp/tmp.gLsjSmW6BS /tmp/tmp.VvVyrRj3q7 ++ return 0 + [[ 2 == \2 ]] + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p7HYW91v9A +++ mktemp ++ local LAST_ERR=/tmp/tmp.FEdth0vi6I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.p7HYW91v9A ++ cat /tmp/tmp.FEdth0vi6I ++ rm /tmp/tmp.p7HYW91v9A /tmp/tmp.FEdth0vi6I ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql /tmp/tmp.xOj8EZf0bc/select-4.sql + desc 'test operator' + set +o xtrace ----------------------------------------------------------------------------------- test operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets operator dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=operator + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.t57FAqWQqN ++ mktemp + local LAST_ERR=/tmp/tmp.T4MGaj14bt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.t57FAqWQqN secret/my-cluster-secrets patched + cat /tmp/tmp.T4MGaj14bt + rm /tmp/tmp.t57FAqWQqN /tmp/tmp.T4MGaj14bt + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7Qph3kGIa4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tOFTnAfHFE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7Qph3kGIa4 ++ cat /tmp/tmp.tOFTnAfHFE ++ rm /tmp/tmp.7Qph3kGIa4 /tmp/tmp.tOFTnAfHFE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wJ4JyPzc6X +++ mktemp ++ local LAST_ERR=/tmp/tmp.y0jkyBqGgI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wJ4JyPzc6X ++ cat /tmp/tmp.y0jkyBqGgI ++ rm /tmp/tmp.wJ4JyPzc6X /tmp/tmp.y0jkyBqGgI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TCqfam20aQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.gnqHEbRDrq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TCqfam20aQ ++ cat /tmp/tmp.gnqHEbRDrq ++ rm /tmp/tmp.TCqfam20aQ /tmp/tmp.gnqHEbRDrq ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EPydxenGKS +++ mktemp ++ local LAST_ERR=/tmp/tmp.uWTnV8Hpuv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EPydxenGKS ++ cat /tmp/tmp.uWTnV8Hpuv ++ rm /tmp/tmp.EPydxenGKS /tmp/tmp.uWTnV8Hpuv ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.uCOrhlYhmB ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.QH5o6JOUTg +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.uCOrhlYhmB +++++ cat /tmp/tmp.QH5o6JOUTg +++++ rm /tmp/tmp.uCOrhlYhmB /tmp/tmp.QH5o6JOUTg +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.7wWAJeQN2g ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Mmf6VBFB7W +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.7wWAJeQN2g +++++ cat /tmp/tmp.Mmf6VBFB7W +++++ rm /tmp/tmp.7wWAJeQN2g /tmp/tmp.Mmf6VBFB7W +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3DAVbCICNe +++ mktemp ++ local LAST_ERR=/tmp/tmp.IWWkdWyV6q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3DAVbCICNe ++ cat /tmp/tmp.IWWkdWyV6q ++ rm /tmp/tmp.3DAVbCICNe /tmp/tmp.IWWkdWyV6q ++ return 0 + [[ 2 == \2 ]] + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qECY0AVdfN +++ mktemp ++ local LAST_ERR=/tmp/tmp.glH7u1ygsM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qECY0AVdfN ++ cat /tmp/tmp.glH7u1ygsM ++ rm /tmp/tmp.qECY0AVdfN /tmp/tmp.glH7u1ygsM ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql /tmp/tmp.xOj8EZf0bc/select-4.sql + desc 'change secret name' + set +o xtrace ----------------------------------------------------------------------------------- change secret name ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.LXjNOxkJhI ++ mktemp + local LAST_ERR=/tmp/tmp.62VmQpW0Gr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LXjNOxkJhI perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.62VmQpW0Gr + rm /tmp/tmp.LXjNOxkJhI /tmp/tmp.62VmQpW0Gr + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iTqpYMMsn6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.db26rcfx94 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iTqpYMMsn6 ++ cat /tmp/tmp.db26rcfx94 ++ rm /tmp/tmp.iTqpYMMsn6 /tmp/tmp.db26rcfx94 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kkUUv0Zwpi +++ mktemp ++ local LAST_ERR=/tmp/tmp.OEWC8mDWZp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kkUUv0Zwpi ++ cat /tmp/tmp.OEWC8mDWZp ++ rm /tmp/tmp.kkUUv0Zwpi /tmp/tmp.OEWC8mDWZp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sP0h0ROI2n +++ mktemp ++ local LAST_ERR=/tmp/tmp.5USQCqup7Q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sP0h0ROI2n ++ cat /tmp/tmp.5USQCqup7Q ++ rm /tmp/tmp.sP0h0ROI2n /tmp/tmp.5USQCqup7Q ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e4zlJE0NMt +++ mktemp ++ local LAST_ERR=/tmp/tmp.ig7q1Hedc1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.e4zlJE0NMt ++ cat /tmp/tmp.ig7q1Hedc1 ++ rm /tmp/tmp.e4zlJE0NMt /tmp/tmp.ig7q1Hedc1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z2hUmSc6SK +++ mktemp ++ local LAST_ERR=/tmp/tmp.Szr3WPV1vu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z2hUmSc6SK ++ cat /tmp/tmp.Szr3WPV1vu ++ rm /tmp/tmp.z2hUmSc6SK /tmp/tmp.Szr3WPV1vu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XS3RfRhWLq +++ mktemp ++ local LAST_ERR=/tmp/tmp.HqPkfxv8up ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XS3RfRhWLq ++ cat /tmp/tmp.HqPkfxv8up ++ rm /tmp/tmp.XS3RfRhWLq /tmp/tmp.HqPkfxv8up ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rAm28GfKK1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.p7fp8fdNKq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rAm28GfKK1 ++ cat /tmp/tmp.p7fp8fdNKq ++ rm /tmp/tmp.rAm28GfKK1 /tmp/tmp.p7fp8fdNKq ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.opiU8WuR49 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ArpLQUKZDB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.opiU8WuR49 ++ cat /tmp/tmp.ArpLQUKZDB ++ rm /tmp/tmp.opiU8WuR49 /tmp/tmp.ArpLQUKZDB ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.cH8vxmLcIJ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Bp7k3Hvu8s +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.cH8vxmLcIJ +++++ cat /tmp/tmp.Bp7k3Hvu8s +++++ rm /tmp/tmp.cH8vxmLcIJ /tmp/tmp.Bp7k3Hvu8s +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.dLfISzI22q ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.OvrmGskoLV +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.dLfISzI22q +++++ cat /tmp/tmp.OvrmGskoLV +++++ rm /tmp/tmp.dLfISzI22q /tmp/tmp.OvrmGskoLV +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nv3C79lMi0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.xKFbdf9P2P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nv3C79lMi0 ++ cat /tmp/tmp.xKFbdf9P2P ++ rm /tmp/tmp.nv3C79lMi0 /tmp/tmp.xKFbdf9P2P ++ return 0 + [[ 2 == \2 ]] + desc 'test new operator' + set +o xtrace ----------------------------------------------------------------------------------- test new operator ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + patch_secret my-cluster-secrets-2 operator dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets-2 + local key=operator + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.pOcHDLizg2 ++ mktemp + local LAST_ERR=/tmp/tmp.DSVJebb2gZ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pOcHDLizg2 secret/my-cluster-secrets-2 patched + cat /tmp/tmp.DSVJebb2gZ + rm /tmp/tmp.pOcHDLizg2 /tmp/tmp.DSVJebb2gZ + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jRSx6dAC5Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.PAxhfadwja ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jRSx6dAC5Q ++ cat /tmp/tmp.PAxhfadwja ++ rm /tmp/tmp.jRSx6dAC5Q /tmp/tmp.PAxhfadwja ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MJl1DmiuoS +++ mktemp ++ local LAST_ERR=/tmp/tmp.z0YeNky7j9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MJl1DmiuoS ++ cat /tmp/tmp.z0YeNky7j9 ++ rm /tmp/tmp.MJl1DmiuoS /tmp/tmp.z0YeNky7j9 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y5uvWZSfzy +++ mktemp ++ local LAST_ERR=/tmp/tmp.32vrxYTm7t ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y5uvWZSfzy ++ cat /tmp/tmp.32vrxYTm7t ++ rm /tmp/tmp.Y5uvWZSfzy /tmp/tmp.32vrxYTm7t ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.t4D4GihY3g ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.98FlxoH9Gg +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.t4D4GihY3g +++++ cat /tmp/tmp.98FlxoH9Gg +++++ rm /tmp/tmp.t4D4GihY3g /tmp/tmp.98FlxoH9Gg +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.ArHhjhrvoC ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.WbVXtcBM5S +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.ArHhjhrvoC +++++ cat /tmp/tmp.WbVXtcBM5S +++++ rm /tmp/tmp.ArHhjhrvoC /tmp/tmp.WbVXtcBM5S +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KnOEPflrf1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tDCfX5PPAq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KnOEPflrf1 ++ cat /tmp/tmp.tDCfX5PPAq ++ rm /tmp/tmp.KnOEPflrf1 /tmp/tmp.tDCfX5PPAq ++ return 0 + [[ 2 == \2 ]] + sleep 20 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N6kT1HGFpz +++ mktemp ++ local LAST_ERR=/tmp/tmp.KV8FyeDj9H ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.N6kT1HGFpz ++ cat /tmp/tmp.KV8FyeDj9H ++ rm /tmp/tmp.N6kT1HGFpz /tmp/tmp.KV8FyeDj9H ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql /tmp/tmp.xOj8EZf0bc/select-4.sql ++ getSecretData my-cluster-secrets-2 root ++ local secretName=my-cluster-secrets-2 ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.SD9jJdmNCt +++ mktemp ++ local LAST_ERR=/tmp/tmp.4xug1cFfIl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SD9jJdmNCt ++ cat /tmp/tmp.4xug1cFfIl ++ rm /tmp/tmp.SD9jJdmNCt /tmp/tmp.4xug1cFfIl ++ return 0 + newpass='p$g?X' + desc 'test new users sync' + set +o xtrace ----------------------------------------------------------------------------------- test new users sync ----------------------------------------------------------------------------------- + run_mysql 'CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''p$g?X'\'';' '-h some-name-pxc -uroot -p'\''p$g?X'\''' + local 'command=CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''p$g?X'\'';' + local 'uri=-h some-name-pxc -uroot -p'\''p$g?X'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aY3bTgcEub +++ mktemp ++ local LAST_ERR=/tmp/tmp.aAj1DSo5Ni ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aY3bTgcEub ++ cat /tmp/tmp.aAj1DSo5Ni ++ rm /tmp/tmp.aY3bTgcEub /tmp/tmp.aAj1DSo5Ni ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + sleep 40 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''p$g?X'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''p$g?X'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''p$g?X'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''p$g?X'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JnwDdlWEDh +++ mktemp ++ local LAST_ERR=/tmp/tmp.990WapmFib ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JnwDdlWEDh ++ cat /tmp/tmp.990WapmFib ++ rm /tmp/tmp.JnwDdlWEDh /tmp/tmp.990WapmFib ++ return 0 + client_pod=pxc-client-7b8b9b895-zq9zb + wait_pod pxc-client-7b8b9b895-zq9zb + local pod=pxc-client-7b8b9b895-zq9zb + local max_retry=480 + local ns= ++ echo pxc-client-7b8b9b895-zq9zb ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-7b8b9b895-zq9zb condition met pxc-client-7b8b9b895-zq9zb.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xOj8EZf0bc/select-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql /tmp/tmp.xOj8EZf0bc/select-4.sql --- /mnt/jenkins/workspace/cloud-pxc-operator_PR-1421/e2e-tests/users/compare/select-4.sql 2023-12-19 08:02:45.152150453 +0000 +++ /tmp/tmp.xOj8EZf0bc/select-4.sql 2023-12-19 10:24:08.792908742 +0000 @@ -1,76 +1,2 @@ -CHARACTER_SETS -CLIENT_STATISTICS -COLLATIONS -COLLATION_CHARACTER_SET_APPLICABILITY -COLUMNS -COLUMN_PRIVILEGES -INDEX_STATISTICS -ENGINES -EVENTS -FILES -GLOBAL_STATUS -GLOBAL_TEMPORARY_TABLES -GLOBAL_VARIABLES -KEY_COLUMN_USAGE -OPTIMIZER_TRACE -PARAMETERS -PARTITIONS -PLUGINS -PROCESSLIST -PROFILING -REFERENTIAL_CONSTRAINTS -ROUTINES -SCHEMATA -SCHEMA_PRIVILEGES -SESSION_STATUS -SESSION_VARIABLES -STATISTICS -TABLES -TABLESPACES -TABLE_CONSTRAINTS -TABLE_PRIVILEGES -TABLE_STATISTICS -TEMPORARY_TABLES -THREAD_STATISTICS -TRIGGERS -USER_PRIVILEGES -USER_STATISTICS -VIEWS -INNODB_CMP -INNODB_CMP_RESET -INNODB_SYS_DATAFILES -XTRADB_READ_VIEW -INNODB_SYS_TABLESTATS -XTRADB_RSEG -INNODB_BUFFER_PAGE -INNODB_TRX -INNODB_CMP_PER_INDEX -INNODB_METRICS -INNODB_FT_DELETED -INNODB_LOCKS -INNODB_LOCK_WAITS -XTRADB_INTERNAL_HASH_TABLES -INNODB_TABLESPACES_ENCRYPTION -INNODB_CMPMEM_RESET -INNODB_SYS_FIELDS -XTRADB_ZIP_DICT -INNODB_TABLESPACES_SCRUBBING -INNODB_TEMP_TABLE_INFO -INNODB_FT_INDEX_TABLE -INNODB_CMPMEM -INNODB_SYS_TABLESPACES -INNODB_CMP_PER_INDEX_RESET -INNODB_SYS_FOREIGN_COLS -INNODB_FT_INDEX_CACHE -INNODB_BUFFER_POOL_STATS -INNODB_FT_BEING_DELETED -INNODB_SYS_FOREIGN -INNODB_BUFFER_PAGE_LRU -INNODB_FT_DEFAULT_STOPWORD -INNODB_SYS_TABLES -INNODB_SYS_COLUMNS -INNODB_FT_CONFIG -XTRADB_ZIP_DICT_COLS -INNODB_SYS_INDEXES -INNODB_SYS_VIRTUAL -INNODB_CHANGED_PAGES +ERROR 1045 (28000): ProxySQL Error: Access denied for user 'testsync'@'10.162.90.52' (using password: YES) +command terminated with exit code 1