Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/logs/custom-users-8-0.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + create_infra custom-users-16422 + local ns=custom-users-16422 + '[' -n pxc-operator ']' + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + kubectl patch pxc -n custom-users-11175 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.eccIYB5EXE ++ mktemp + local LAST_ERR=/tmp/tmp.mfUOwasMSO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eccIYB5EXE perconaxtradbcluster.pxc.percona.com "some-name" deleted from custom-users-11175 namespace + cat /tmp/tmp.mfUOwasMSO + rm /tmp/tmp.eccIYB5EXE /tmp/tmp.mfUOwasMSO + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.jsaPTtAmOy ++ mktemp + local LAST_ERR=/tmp/tmp.A6MeBlb7ee + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jsaPTtAmOy No resources found + cat /tmp/tmp.A6MeBlb7ee + rm /tmp/tmp.jsaPTtAmOy /tmp/tmp.A6MeBlb7ee + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.Ggi83QBJZc ++ mktemp + local LAST_ERR=/tmp/tmp.51qfmRGNGt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Ggi83QBJZc No resources found + cat /tmp/tmp.51qfmRGNGt + rm /tmp/tmp.Ggi83QBJZc /tmp/tmp.51qfmRGNGt + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ awk '-F ' '{print $2}' ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + xargs kubectl delete ns + awk '{print$1}' + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.Da6uyVYdVc ++ mktemp + local LAST_ERR=/tmp/tmp.TtWqUMlBTQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns ++ mktemp + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + local LAST_OUT=/tmp/tmp.tWFHhW1HZl ++ mktemp + local LAST_ERR=/tmp/tmp.tw8vQzjPmH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Da6uyVYdVc + cat /tmp/tmp.TtWqUMlBTQ + rm /tmp/tmp.Da6uyVYdVc /tmp/tmp.TtWqUMlBTQ + return 0 namespace "custom-users-11175" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tWFHhW1HZl namespace "pxc-operator" deleted + cat /tmp/tmp.tw8vQzjPmH + rm /tmp/tmp.tWFHhW1HZl /tmp/tmp.tw8vQzjPmH + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.oXMoNEttzp ++ mktemp + local LAST_ERR=/tmp/tmp.TxmSxOpOT0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oXMoNEttzp namespace/pxc-operator created + cat /tmp/tmp.TxmSxOpOT0 + rm /tmp/tmp.oXMoNEttzp /tmp/tmp.TxmSxOpOT0 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.vmDDhb3QIH +++ mktemp ++ local LAST_ERR=/tmp/tmp.G6Qrilz0QK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vmDDhb3QIH ++ cat /tmp/tmp.G6Qrilz0QK ++ rm /tmp/tmp.vmDDhb3QIH /tmp/tmp.G6Qrilz0QK ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2228-1e2eac2d-3-cluster4 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.BTydSCYzfY ++ mktemp + local LAST_ERR=/tmp/tmp.0U5eUZTfPI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2228-1e2eac2d-3-cluster4 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BTydSCYzfY Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2228-1e2eac2d-3-cluster4" modified. + cat /tmp/tmp.0U5eUZTfPI + rm /tmp/tmp.BTydSCYzfY /tmp/tmp.0U5eUZTfPI + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.EgOIhO9Qp7 ++ mktemp + local LAST_ERR=/tmp/tmp.crFt7JWqKW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EgOIhO9Qp7 customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.crFt7JWqKW + rm /tmp/tmp.EgOIhO9Qp7 /tmp/tmp.crFt7JWqKW + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + kubectl_bin apply -f - + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + local LAST_OUT=/tmp/tmp.B7TE6KoM9H ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/deploy/cw-rbac.yaml + local LAST_ERR=/tmp/tmp.DIv6h6cjXb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.B7TE6KoM9H clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.DIv6h6cjXb + rm /tmp/tmp.B7TE6KoM9H /tmp/tmp.DIv6h6cjXb + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2228-1e2eac2d^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ln6XWVK0Ez ++ mktemp + local LAST_ERR=/tmp/tmp.IMXCj3IX2i + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ln6XWVK0Ez deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.IMXCj3IX2i + rm /tmp/tmp.ln6XWVK0Ez /tmp/tmp.IMXCj3IX2i + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.P7riztxSM8 ++ mktemp + local LAST_ERR=/tmp/tmp.NVFTri2ozT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.P7riztxSM8 pod/percona-xtradb-cluster-operator-5c9f99c4c6-pww79 condition met + cat /tmp/tmp.NVFTri2ozT + rm /tmp/tmp.P7riztxSM8 /tmp/tmp.NVFTri2ozT + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.RFaB4isVAV +++ mktemp ++ local LAST_ERR=/tmp/tmp.KDYBFLCSrF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RFaB4isVAV ++ cat /tmp/tmp.KDYBFLCSrF ++ rm /tmp/tmp.RFaB4isVAV /tmp/tmp.KDYBFLCSrF ++ return 0 + wait_pod percona-xtradb-cluster-operator-5c9f99c4c6-pww79 480 pxc-operator + local pod=percona-xtradb-cluster-operator-5c9f99c4c6-pww79 + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-5c9f99c4c6-pww79 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-5c9f99c4c6-pww79 condition met waiting for pod/percona-xtradb-cluster-operator-5c9f99c4c6-pww79 to become Ready.Ok + sleep 3 + create_namespace custom-users-16422 + local namespace=custom-users-16422 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces custom-users-16422' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces custom-users-16422 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace custom-users-16422 ++ mktemp + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp + local LAST_OUT=/tmp/tmp.T4qRRhZjMu + local LAST_OUT=/tmp/tmp.Ej2qgj1b9o ++ mktemp + local LAST_ERR=/tmp/tmp.eQ91hHwZA9 + local exit_status=0 ++ mktemp + local LAST_ERR=/tmp/tmp.IfoMWzSs4i + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace custom-users-16422 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Ej2qgj1b9o + cat /tmp/tmp.IfoMWzSs4i + rm /tmp/tmp.Ej2qgj1b9o /tmp/tmp.IfoMWzSs4i + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace custom-users-16422 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace custom-users-16422 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.T4qRRhZjMu + cat /tmp/tmp.eQ91hHwZA9 Error from server (NotFound): namespaces "custom-users-16422" not found + rm /tmp/tmp.T4qRRhZjMu /tmp/tmp.eQ91hHwZA9 + return 1 + : + wait_for_delete namespace/custom-users-16422 + local res=namespace/custom-users-16422 + echo -n 'waiting for namespace/custom-users-16422 to be deleted' waiting for namespace/custom-users-16422 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "custom-users-16422" not found + desc 'create namespace custom-users-16422' + set +o xtrace ----------------------------------------------------------------------------------- create namespace custom-users-16422 ----------------------------------------------------------------------------------- + kubectl_bin create namespace custom-users-16422 ++ mktemp + local LAST_OUT=/tmp/tmp.QrbBvd0SDk ++ mktemp + local LAST_ERR=/tmp/tmp.MS6LwN1tOO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace custom-users-16422 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QrbBvd0SDk namespace/custom-users-16422 created + cat /tmp/tmp.MS6LwN1tOO + rm /tmp/tmp.QrbBvd0SDk /tmp/tmp.MS6LwN1tOO + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZjMFJ1bnPj +++ mktemp ++ local LAST_ERR=/tmp/tmp.voTR2pCIj9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZjMFJ1bnPj ++ cat /tmp/tmp.voTR2pCIj9 ++ rm /tmp/tmp.ZjMFJ1bnPj /tmp/tmp.voTR2pCIj9 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2228-1e2eac2d-3-cluster4 --namespace=custom-users-16422 ++ mktemp + local LAST_OUT=/tmp/tmp.ubiyk9hB33 ++ mktemp + local LAST_ERR=/tmp/tmp.mHaLhmPq1h + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2228-1e2eac2d-3-cluster4 --namespace=custom-users-16422 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ubiyk9hB33 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2228-1e2eac2d-3-cluster4" modified. + cat /tmp/tmp.mHaLhmPq1h + rm /tmp/tmp.ubiyk9hB33 /tmp/tmp.mHaLhmPq1h + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.xQZCxKNeBR ++ mktemp + local LAST_ERR=/tmp/tmp.XJMtFBZwzf + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xQZCxKNeBR secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.XJMtFBZwzf + rm /tmp/tmp.xQZCxKNeBR /tmp/tmp.XJMtFBZwzf + return 0 + desc 'create PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- + cluster=some-name + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/conf/user-secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.DD6Ny4ihEU ++ mktemp + local LAST_ERR=/tmp/tmp.NFzzFhAod6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/conf/user-secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.DD6Ny4ihEU secret/user-secrets created + cat /tmp/tmp.NFzzFhAod6 + rm /tmp/tmp.DD6Ny4ihEU /tmp/tmp.NFzzFhAod6 + return 0 + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.q2OJ9h9b0O ++ mktemp + local LAST_ERR=/tmp/tmp.2ZEiFiHrcN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.q2OJ9h9b0O secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.2ZEiFiHrcN + rm /tmp/tmp.q2OJ9h9b0O /tmp/tmp.2ZEiFiHrcN + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/client.yml + kubectl_bin apply -f - + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2228-1e2eac2d#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.custom-users-16422~ + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + local LAST_OUT=/tmp/tmp.rSdNmtR0FY + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + local LAST_ERR=/tmp/tmp.jaKxuk6PoP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rSdNmtR0FY deployment.apps/pxc-client created + cat /tmp/tmp.jaKxuk6PoP + rm /tmp/tmp.rSdNmtR0FY /tmp/tmp.jaKxuk6PoP + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + kubectl_bin apply -f - + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2228-1e2eac2d#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + /usr/bin/sed -e 's#apply:.*#apply: Never#' + local LAST_OUT=/tmp/tmp.duejZCnn7R + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/conf/some-name.yml ++ mktemp + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.custom-users-16422~ + local LAST_ERR=/tmp/tmp.Zmlugj7mdj + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.duejZCnn7R perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.Zmlugj7mdj + rm /tmp/tmp.duejZCnn7R /tmp/tmp.Zmlugj7mdj + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Zya2z8FeEw ++++ mktemp +++ local LAST_ERR=/tmp/tmp.QA2wrb4rI3 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Zya2z8FeEw +++ cat /tmp/tmp.QA2wrb4rI3 +++ rm /tmp/tmp.Zya2z8FeEw /tmp/tmp.QA2wrb4rI3 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-haproxy ++ return + local proxy=some-name-haproxy + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-16422 ++ mktemp + local LAST_OUT=/tmp/tmp.iQMs3JJqwp ++ mktemp + local LAST_ERR=/tmp/tmp.EfxkXXQXvv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-16422 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-16422 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-16422 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.iQMs3JJqwp + cat /tmp/tmp.EfxkXXQXvv error: no matching resources found + rm /tmp/tmp.iQMs3JJqwp /tmp/tmp.EfxkXXQXvv + return 1 + true + wait_for_running some-name-haproxy 1 + local name=some-name-haproxy + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-haproxy-0 480 + local pod=some-name-haproxy-0 + local max_retry=480 + local ns= ++ echo some-name-haproxy-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/some-name-haproxy-0 condition met waiting for pod/some-name-haproxy-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ base64 --decode ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U1W6WAdq5D +++ mktemp ++ local LAST_ERR=/tmp/tmp.3F1937GS8Q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.U1W6WAdq5D ++ cat /tmp/tmp.3F1937GS8Q ++ rm /tmp/tmp.U1W6WAdq5D /tmp/tmp.3F1937GS8Q ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1yKklfCzzL +++ mktemp ++ local LAST_ERR=/tmp/tmp.xs6ShfQO00 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1yKklfCzzL ++ cat /tmp/tmp.xs6ShfQO00 ++ rm /tmp/tmp.1yKklfCzzL /tmp/tmp.xs6ShfQO00 ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WhCs0PtdK3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.60MFPCt7EU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WhCs0PtdK3 ++ cat /tmp/tmp.60MFPCt7EU ++ rm /tmp/tmp.WhCs0PtdK3 /tmp/tmp.60MFPCt7EU ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pBw4sJcrGu +++ mktemp ++ local LAST_ERR=/tmp/tmp.8hr3CwnTRD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pBw4sJcrGu ++ cat /tmp/tmp.8hr3CwnTRD ++ rm /tmp/tmp.pBw4sJcrGu /tmp/tmp.8hr3CwnTRD ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1.sql /tmp/tmp.GW4coCzP8f/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RS2cfDLPwh +++ mktemp ++ local LAST_ERR=/tmp/tmp.hjxTYrkF7S ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RS2cfDLPwh ++ cat /tmp/tmp.hjxTYrkF7S ++ rm /tmp/tmp.RS2cfDLPwh /tmp/tmp.hjxTYrkF7S ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1.sql /tmp/tmp.GW4coCzP8f/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LutUvzH6dD +++ mktemp ++ local LAST_ERR=/tmp/tmp.Upc1aMbxEX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LutUvzH6dD ++ cat /tmp/tmp.Upc1aMbxEX ++ rm /tmp/tmp.LutUvzH6dD /tmp/tmp.Upc1aMbxEX ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/select-1.sql /tmp/tmp.GW4coCzP8f/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ egrep -o 'early-plugin-load=keyring_\w+.so' ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SsDhBDors4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.YhMFxPCFKU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SsDhBDors4 ++ cat /tmp/tmp.YhMFxPCFKU Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.SsDhBDors4 /tmp/tmp.YhMFxPCFKU ++ return 0 + '[' '' ']' + desc 'check users created on cluster creation' + set +o xtrace ----------------------------------------------------------------------------------- check users created on cluster creation ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-haproxy -uuser-one -ptestpass' + local 'uri=-h some-name-haproxy -uuser-one -ptestpass' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-haproxy -uuser-one -ptestpass + local user=user-one + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-one -ptestpass' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-one -ptestpass' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WZdEfC7FkO +++ mktemp ++ local LAST_ERR=/tmp/tmp.3fTrNQNprH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WZdEfC7FkO ++ cat /tmp/tmp.3fTrNQNprH ++ rm /tmp/tmp.WZdEfC7FkO /tmp/tmp.3fTrNQNprH ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one.sql /tmp/tmp.GW4coCzP8f/user-one.sql + compare_mysql_cmd user-one-1 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fcAw41yJSd +++ mktemp ++ local LAST_ERR=/tmp/tmp.X0S3FBqtx6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fcAw41yJSd ++ cat /tmp/tmp.X0S3FBqtx6 ++ rm /tmp/tmp.fcAw41yJSd /tmp/tmp.X0S3FBqtx6 ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-kcg2v + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/user-one-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-1.sql /tmp/tmp.GW4coCzP8f/user-one-1.sql + compare_mysql_cmd user-one-2 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-2 + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i0z6plvEbZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.5THDY7L9Zm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.i0z6plvEbZ ++ cat /tmp/tmp.5THDY7L9Zm ++ rm /tmp/tmp.i0z6plvEbZ /tmp/tmp.5THDY7L9Zm ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/user-one-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-2.sql /tmp/tmp.GW4coCzP8f/user-one-2.sql + compare_mysql_user '-h some-name-haproxy -uuser-two -ptestpass3' + local 'uri=-h some-name-haproxy -uuser-two -ptestpass3' + local postfix= ++ echo -h some-name-haproxy -uuser-two -ptestpass3 ++ sed -e 's/.*-u//; s/ .*//' + local user=user-two + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-two -ptestpass3' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-two -ptestpass3' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IdHsqF8WSp +++ mktemp ++ local LAST_ERR=/tmp/tmp.nQzis9kUUk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IdHsqF8WSp ++ cat /tmp/tmp.nQzis9kUUk ++ rm /tmp/tmp.IdHsqF8WSp /tmp/tmp.nQzis9kUUk ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two.sql /tmp/tmp.GW4coCzP8f/user-two.sql + compare_mysql_cmd user-two-1 'SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-two-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0SavEg5fFG +++ mktemp ++ local LAST_ERR=/tmp/tmp.VNTS1WUBt4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0SavEg5fFG ++ cat /tmp/tmp.VNTS1WUBt4 ++ rm /tmp/tmp.0SavEg5fFG /tmp/tmp.VNTS1WUBt4 ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-kcg2v ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/user-two-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two-1.sql /tmp/tmp.GW4coCzP8f/user-two-1.sql + generatedUserSecret=some-name-custom-user-secret ++ kubectl_bin get secret some-name-custom-user-secret -o 'jsonpath={.data.user-three}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KUU5UtGK4R +++ mktemp ++ local LAST_ERR=/tmp/tmp.rXDUUEKayD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secret some-name-custom-user-secret -o 'jsonpath={.data.user-three}' ++ base64 -d ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KUU5UtGK4R ++ cat /tmp/tmp.rXDUUEKayD ++ rm /tmp/tmp.KUU5UtGK4R /tmp/tmp.rXDUUEKayD ++ return 0 + userThreePass='g>VbL&zO*wc~N>tac5Z' + compare_mysql_user '-h some-name-haproxy -uuser-three -p'\''g>VbL&zO*wc~N>tac5Z'\''' + local 'uri=-h some-name-haproxy -uuser-three -p'\''g>VbL&zO*wc~N>tac5Z'\''' + local postfix= ++ echo -h some-name-haproxy -uuser-three '-p'\''g>VbL&zO*wc~N>tac5Z'\''' ++ sed -e 's/.*-u//; s/ .*//' + local user=user-three + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-three.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-three-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-three -p'\''g>VbL&zO*wc~N>tac5Z'\''' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-three -p'\''g>VbL&zO*wc~N>tac5Z'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ local LAST_OUT=/tmp/tmp.JsYyHdpE8w +++ mktemp ++ local LAST_ERR=/tmp/tmp.aee1X4lOlU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JsYyHdpE8w ++ cat /tmp/tmp.aee1X4lOlU ++ rm /tmp/tmp.JsYyHdpE8w /tmp/tmp.aee1X4lOlU ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-three.sql /tmp/tmp.GW4coCzP8f/user-three.sql + compare_mysql_cmd user-three-1 'SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-three-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-three-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-three-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.E7P9OaC4KW +++ mktemp ++ local LAST_ERR=/tmp/tmp.y7nKmu1GoQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.E7P9OaC4KW ++ cat /tmp/tmp.y7nKmu1GoQ ++ rm /tmp/tmp.E7P9OaC4KW /tmp/tmp.y7nKmu1GoQ ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/user-three-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-three-1.sql /tmp/tmp.GW4coCzP8f/user-three-1.sql + desc 'check password change' + set +o xtrace ----------------------------------------------------------------------------------- check password change ----------------------------------------------------------------------------------- + kubectl_bin patch secret user-secrets '-p={"stringData":{"pwd-key-one": "new-password"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.rPloXFeaq1 ++ mktemp + local LAST_ERR=/tmp/tmp.8xdYbVxXR9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret user-secrets '-p={"stringData":{"pwd-key-one": "new-password"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rPloXFeaq1 secret/user-secrets patched + cat /tmp/tmp.8xdYbVxXR9 + rm /tmp/tmp.rPloXFeaq1 /tmp/tmp.8xdYbVxXR9 + return 0 + sleep 15 + compare_mysql_user '-h some-name-haproxy -uuser-one -pnew-password' + local 'uri=-h some-name-haproxy -uuser-one -pnew-password' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-haproxy -uuser-one -pnew-password + local user=user-one + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-one -pnew-password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-one -pnew-password' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zdF2qc08IM +++ mktemp ++ local LAST_ERR=/tmp/tmp.3ua9M25Oym ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zdF2qc08IM ++ cat /tmp/tmp.3ua9M25Oym ++ rm /tmp/tmp.zdF2qc08IM /tmp/tmp.3ua9M25Oym ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-kcg2v ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one.sql /tmp/tmp.GW4coCzP8f/user-one.sql + compare_mysql_cmd user-one-1 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OZ71JBXeaR +++ mktemp ++ local LAST_ERR=/tmp/tmp.duA6QLvJT2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OZ71JBXeaR ++ cat /tmp/tmp.duA6QLvJT2 ++ rm /tmp/tmp.OZ71JBXeaR /tmp/tmp.duA6QLvJT2 ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-kcg2v ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/user-one-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-1.sql /tmp/tmp.GW4coCzP8f/user-one-1.sql + compare_mysql_cmd user-one-2 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-2 + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8YljeM70RO +++ mktemp ++ local LAST_ERR=/tmp/tmp.YEjciA3NcE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8YljeM70RO ++ cat /tmp/tmp.YEjciA3NcE ++ rm /tmp/tmp.8YljeM70RO /tmp/tmp.YEjciA3NcE ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.GW4coCzP8f/user-one-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-one-2.sql /tmp/tmp.GW4coCzP8f/user-one-2.sql + desc 'check removing secretPasswordRef from user generates user password stored in generated custom-user-secret' + set +o xtrace ----------------------------------------------------------------------------------- check removing secretPasswordRef from user generates user password stored in generated custom-user-secret ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=json '-p=[{"op": "replace", "path": "/spec/users/1/passwordSecretRef", "value": null}]' ++ mktemp + local LAST_OUT=/tmp/tmp.1m6pFt8wmb ++ mktemp + local LAST_ERR=/tmp/tmp.OqF7y4YndF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=json '-p=[{"op": "replace", "path": "/spec/users/1/passwordSecretRef", "value": null}]' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1m6pFt8wmb perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.OqF7y4YndF + rm /tmp/tmp.1m6pFt8wmb /tmp/tmp.OqF7y4YndF + return 0 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xw3xLEI27b +++ mktemp ++ local LAST_ERR=/tmp/tmp.AZ3D7fWsan ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xw3xLEI27b ++ cat /tmp/tmp.AZ3D7fWsan ++ rm /tmp/tmp.Xw3xLEI27b /tmp/tmp.AZ3D7fWsan ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lp6egoOgym +++ mktemp ++ local LAST_ERR=/tmp/tmp.mXCwlc7NQh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lp6egoOgym ++ cat /tmp/tmp.mXCwlc7NQh ++ rm /tmp/tmp.lp6egoOgym /tmp/tmp.mXCwlc7NQh ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.elNAM37S8i ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.3gO6e1tmNC +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.elNAM37S8i +++++ cat /tmp/tmp.3gO6e1tmNC +++++ rm /tmp/tmp.elNAM37S8i /tmp/tmp.3gO6e1tmNC +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.11Km8OaNsd +++ mktemp ++ local LAST_ERR=/tmp/tmp.l7onSdLBz0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.11Km8OaNsd ++ cat /tmp/tmp.l7onSdLBz0 ++ rm /tmp/tmp.11Km8OaNsd /tmp/tmp.l7onSdLBz0 ++ return 0 + [[ 3 == \3 ]] + echo ++ kubectl_bin get secret some-name-custom-user-secret -o 'jsonpath={.data.user-two}' ++ base64 -d +++ mktemp ++ local LAST_OUT=/tmp/tmp.LBwMbN8k4W +++ mktemp ++ local LAST_ERR=/tmp/tmp.i7hXkWRf2B ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secret some-name-custom-user-secret -o 'jsonpath={.data.user-two}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LBwMbN8k4W ++ cat /tmp/tmp.i7hXkWRf2B ++ rm /tmp/tmp.LBwMbN8k4W /tmp/tmp.i7hXkWRf2B ++ return 0 + userTwoPass==KS%nGmByC82ogd3NF + compare_mysql_user '-h some-name-haproxy -uuser-two -p'\''=KS%nGmByC82ogd3NF'\''' + local 'uri=-h some-name-haproxy -uuser-two -p'\''=KS%nGmByC82ogd3NF'\''' + local postfix= ++ echo -h some-name-haproxy -uuser-two '-p'\''=KS%nGmByC82ogd3NF'\''' ++ sed -e 's/.*-u//; s/ .*//' + local user=user-two + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-two -p'\''=KS%nGmByC82ogd3NF'\''' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-two -p'\''=KS%nGmByC82ogd3NF'\''' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.czM4pnmsVH +++ mktemp ++ local LAST_ERR=/tmp/tmp.PyIQWYlkAP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.czM4pnmsVH ++ cat /tmp/tmp.PyIQWYlkAP ++ rm /tmp/tmp.czM4pnmsVH /tmp/tmp.PyIQWYlkAP ++ return 0 + client_pod=pxc-client-59944c5bbf-kcg2v + wait_pod pxc-client-59944c5bbf-kcg2v + local pod=pxc-client-59944c5bbf-kcg2v + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-kcg2v ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-kcg2v condition met waiting for pod/pxc-client-59944c5bbf-kcg2v to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two.sql /tmp/tmp.GW4coCzP8f/user-two.sql --- /mnt/jenkins/workspace/cloud-pxc-operator_PR-2228/e2e-tests/custom-users/compare/user-two.sql 2025-11-03 12:25:39.353411600 +0000 +++ /tmp/tmp.GW4coCzP8f/user-two.sql 2025-11-03 12:58:10.365711495 +0000 @@ -1 +1,2 @@ -GRANT INSERT, UPDATE ON *.* TO `user-two`@`%` +ERROR 1045 (28000): Access denied for user 'user-two'@ (using password: YES) +command terminated with exit code 1