Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/logs/custom-users-8-0.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + create_infra custom-users-20369 + local ns=custom-users-20369 + '[' -n pxc-operator ']' + grep -v NAMESPACE + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n custom-users-4598 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.uIWkSFZBKf ++ mktemp + local LAST_ERR=/tmp/tmp.1Ws5OpZ1Xf + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uIWkSFZBKf perconaxtradbcluster.pxc.percona.com "some-name" deleted from custom-users-4598 namespace + cat /tmp/tmp.1Ws5OpZ1Xf + rm /tmp/tmp.uIWkSFZBKf /tmp/tmp.1Ws5OpZ1Xf + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.dC4kXC5FxS ++ mktemp + local LAST_ERR=/tmp/tmp.RM8HzTHLMw + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dC4kXC5FxS No resources found + cat /tmp/tmp.RM8HzTHLMw + rm /tmp/tmp.dC4kXC5FxS /tmp/tmp.RM8HzTHLMw + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.0J47OVHCWF ++ mktemp + local LAST_ERR=/tmp/tmp.LbrNpD6bBt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0J47OVHCWF No resources found + cat /tmp/tmp.LbrNpD6bBt + rm /tmp/tmp.0J47OVHCWF /tmp/tmp.LbrNpD6bBt + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// ++ tail -n1 ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrolebinding ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + kubectl_bin get ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.GINWvdr7L4 + local LAST_OUT=/tmp/tmp.7gzjdBXGSK ++ mktemp + local LAST_ERR=/tmp/tmp.qbGOI2JGtd + local exit_status=0 ++ mktemp + awk '{print$1}' + local LAST_ERR=/tmp/tmp.8jIiskrxVQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + xargs kubectl delete ns ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.GINWvdr7L4 + cat /tmp/tmp.qbGOI2JGtd + rm /tmp/tmp.GINWvdr7L4 /tmp/tmp.qbGOI2JGtd + return 0 namespace "custom-users-4598" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.7gzjdBXGSK namespace "pxc-operator" deleted + cat /tmp/tmp.8jIiskrxVQ + rm /tmp/tmp.7gzjdBXGSK /tmp/tmp.8jIiskrxVQ + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.inq0Gm595Y ++ mktemp + local LAST_ERR=/tmp/tmp.PubT3fB6I2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.inq0Gm595Y namespace/pxc-operator created + cat /tmp/tmp.PubT3fB6I2 + rm /tmp/tmp.inq0Gm595Y /tmp/tmp.PubT3fB6I2 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.xx90nL3dVz +++ mktemp ++ local LAST_ERR=/tmp/tmp.D5TtCiTbzb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xx90nL3dVz ++ cat /tmp/tmp.D5TtCiTbzb ++ rm /tmp/tmp.xx90nL3dVz /tmp/tmp.D5TtCiTbzb ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster4 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.HtLzKDucWq ++ mktemp + local LAST_ERR=/tmp/tmp.8YUOngdXrS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster4 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HtLzKDucWq Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster4" modified. + cat /tmp/tmp.8YUOngdXrS + rm /tmp/tmp.HtLzKDucWq /tmp/tmp.8YUOngdXrS + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.GNV0GgYhLX ++ mktemp + local LAST_ERR=/tmp/tmp.Dr0r6fwkf2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.GNV0GgYhLX customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.Dr0r6fwkf2 + rm /tmp/tmp.GNV0GgYhLX /tmp/tmp.Dr0r6fwkf2 + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + local LAST_OUT=/tmp/tmp.DagmX8rdh1 ++ mktemp + local LAST_ERR=/tmp/tmp.Kc60sVpEpb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.DagmX8rdh1 clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.Kc60sVpEpb + rm /tmp/tmp.DagmX8rdh1 /tmp/tmp.Kc60sVpEpb + return 0 + kubectl_bin apply -f - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2207-89209ce1^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - ++ mktemp + local LAST_OUT=/tmp/tmp.60KeI0i4AS ++ mktemp + local LAST_ERR=/tmp/tmp.a7Sy6Z4iH4 + local exit_status=0 ++ seq 0 2 + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/cw-operator.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.60KeI0i4AS deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.a7Sy6Z4iH4 + rm /tmp/tmp.60KeI0i4AS /tmp/tmp.a7Sy6Z4iH4 + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.8lUcHxpwDe ++ mktemp + local LAST_ERR=/tmp/tmp.c5c7Wg7X6Y + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.8lUcHxpwDe pod/percona-xtradb-cluster-operator-58dd9fd94c-c49gt condition met + cat /tmp/tmp.c5c7Wg7X6Y + rm /tmp/tmp.8lUcHxpwDe /tmp/tmp.c5c7Wg7X6Y + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ grep -c percona-xtradb-cluster-operator +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.rpQd7mwZGI +++ mktemp ++ local LAST_ERR=/tmp/tmp.mNgQXPOdFc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rpQd7mwZGI ++ cat /tmp/tmp.mNgQXPOdFc ++ rm /tmp/tmp.rpQd7mwZGI /tmp/tmp.mNgQXPOdFc ++ return 0 + wait_pod percona-xtradb-cluster-operator-58dd9fd94c-c49gt 480 pxc-operator + local pod=percona-xtradb-cluster-operator-58dd9fd94c-c49gt + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-58dd9fd94c-c49gt ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-58dd9fd94c-c49gt condition met waiting for pod/percona-xtradb-cluster-operator-58dd9fd94c-c49gt to become Ready.Ok + sleep 3 + create_namespace custom-users-20369 + local namespace=custom-users-20369 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces custom-users-20369' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces custom-users-20369 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace custom-users-20369 ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.wUqjXsdi5A + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp + local LAST_ERR=/tmp/tmp.dgDSfFvpud + local exit_status=0 ++ seq 0 2 + xargs kubectl delete ns + kubectl_bin get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace custom-users-20369 ++ mktemp + local LAST_OUT=/tmp/tmp.Su4puo9aIj ++ mktemp + local LAST_ERR=/tmp/tmp.kHCCcCxgT6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace custom-users-20369 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Su4puo9aIj + cat /tmp/tmp.kHCCcCxgT6 + rm /tmp/tmp.Su4puo9aIj /tmp/tmp.kHCCcCxgT6 + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace custom-users-20369 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.wUqjXsdi5A + cat /tmp/tmp.dgDSfFvpud Error from server (NotFound): namespaces "custom-users-20369" not found + rm /tmp/tmp.wUqjXsdi5A /tmp/tmp.dgDSfFvpud + return 1 + : + wait_for_delete namespace/custom-users-20369 + local res=namespace/custom-users-20369 + echo -n 'waiting for namespace/custom-users-20369 to be deleted' waiting for namespace/custom-users-20369 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "custom-users-20369" not found + desc 'create namespace custom-users-20369' + set +o xtrace ----------------------------------------------------------------------------------- create namespace custom-users-20369 ----------------------------------------------------------------------------------- + kubectl_bin create namespace custom-users-20369 ++ mktemp + local LAST_OUT=/tmp/tmp.kVparEV7vw ++ mktemp + local LAST_ERR=/tmp/tmp.RkLExOiUpK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace custom-users-20369 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.kVparEV7vw namespace/custom-users-20369 created + cat /tmp/tmp.RkLExOiUpK + rm /tmp/tmp.kVparEV7vw /tmp/tmp.RkLExOiUpK + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.iTSNWCxdjK +++ mktemp ++ local LAST_ERR=/tmp/tmp.itSFX2ltXK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iTSNWCxdjK ++ cat /tmp/tmp.itSFX2ltXK ++ rm /tmp/tmp.iTSNWCxdjK /tmp/tmp.itSFX2ltXK ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster4 --namespace=custom-users-20369 ++ mktemp + local LAST_OUT=/tmp/tmp.gtndo46Rr3 ++ mktemp + local LAST_ERR=/tmp/tmp.zlIcP1S59l + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster4 --namespace=custom-users-20369 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.gtndo46Rr3 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster4" modified. + cat /tmp/tmp.zlIcP1S59l + rm /tmp/tmp.gtndo46Rr3 /tmp/tmp.zlIcP1S59l + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.vu7SPQs1y5 ++ mktemp + local LAST_ERR=/tmp/tmp.gmlYVsCYFn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vu7SPQs1y5 secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.gmlYVsCYFn + rm /tmp/tmp.vu7SPQs1y5 /tmp/tmp.gmlYVsCYFn + return 0 + desc 'create PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- + cluster=some-name + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/conf/user-secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.6FMT4iGTYH ++ mktemp + local LAST_ERR=/tmp/tmp.lSizuc0MPX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/conf/user-secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6FMT4iGTYH secret/user-secrets created + cat /tmp/tmp.lSizuc0MPX + rm /tmp/tmp.6FMT4iGTYH /tmp/tmp.lSizuc0MPX + return 0 + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.1lWRIdaj28 ++ mktemp + local LAST_ERR=/tmp/tmp.ZjVAhW92hS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1lWRIdaj28 secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.ZjVAhW92hS + rm /tmp/tmp.1lWRIdaj28 /tmp/tmp.ZjVAhW92hS + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.custom-users-20369~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2207-89209ce1#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + local LAST_OUT=/tmp/tmp.dCdCvW3mKn ++ mktemp + local LAST_ERR=/tmp/tmp.1QmQX9Rlmg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dCdCvW3mKn deployment.apps/pxc-client created + cat /tmp/tmp.1QmQX9Rlmg + rm /tmp/tmp.dCdCvW3mKn /tmp/tmp.1QmQX9Rlmg + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/conf/some-name.yml + /usr/bin/sed -e s~minio-service.#namespace~minio-service.custom-users-20369~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_OUT=/tmp/tmp.oFwdnL9L80 + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2207-89209ce1#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/conf/some-name.yml ++ mktemp + local LAST_ERR=/tmp/tmp.gFBkYkxq22 + local exit_status=0 + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oFwdnL9L80 perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.gFBkYkxq22 + rm /tmp/tmp.oFwdnL9L80 /tmp/tmp.gFBkYkxq22 + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.IH86aFGaVx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.PpZbPJpgnr +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.IH86aFGaVx +++ cat /tmp/tmp.PpZbPJpgnr +++ rm /tmp/tmp.IH86aFGaVx /tmp/tmp.PpZbPJpgnr +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-haproxy ++ return + local proxy=some-name-haproxy + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-20369 ++ mktemp + local LAST_OUT=/tmp/tmp.w5qbCgDkba ++ mktemp + local LAST_ERR=/tmp/tmp.9xaThIFTZt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-20369 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-20369 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n custom-users-20369 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.w5qbCgDkba + cat /tmp/tmp.9xaThIFTZt error: no matching resources found + rm /tmp/tmp.w5qbCgDkba /tmp/tmp.9xaThIFTZt + return 1 + true + wait_for_running some-name-haproxy 1 + local name=some-name-haproxy + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-haproxy-0 480 + local pod=some-name-haproxy-0 + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-haproxy-0 + local container= + set +o xtrace pod/some-name-haproxy-0 condition met waiting for pod/some-name-haproxy-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ base64 --decode ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8o2Nu274po +++ mktemp ++ local LAST_ERR=/tmp/tmp.pELXoKD3AN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8o2Nu274po ++ cat /tmp/tmp.pELXoKD3AN ++ rm /tmp/tmp.8o2Nu274po /tmp/tmp.pELXoKD3AN ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WkXkI6Kj1i +++ mktemp ++ local LAST_ERR=/tmp/tmp.niYsCQltRI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WkXkI6Kj1i ++ cat /tmp/tmp.niYsCQltRI ++ rm /tmp/tmp.WkXkI6Kj1i /tmp/tmp.niYsCQltRI ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-s5vdf + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-haproxy -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iy03Ym4jRd +++ mktemp ++ local LAST_ERR=/tmp/tmp.H8CWvTlXh4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iy03Ym4jRd ++ cat /tmp/tmp.H8CWvTlXh4 ++ rm /tmp/tmp.iy03Ym4jRd /tmp/tmp.H8CWvTlXh4 ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-s5vdf + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B2uZV9ex7C +++ mktemp ++ local LAST_ERR=/tmp/tmp.mzzaD0k2TO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B2uZV9ex7C ++ cat /tmp/tmp.mzzaD0k2TO ++ rm /tmp/tmp.B2uZV9ex7C /tmp/tmp.mzzaD0k2TO ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-s5vdf + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1.sql /tmp/tmp.U7OmJ8sNiE/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6JWqRRtPrK +++ mktemp ++ local LAST_ERR=/tmp/tmp.xpJdSq7txo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6JWqRRtPrK ++ cat /tmp/tmp.xpJdSq7txo ++ rm /tmp/tmp.6JWqRRtPrK /tmp/tmp.xpJdSq7txo ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1.sql /tmp/tmp.U7OmJ8sNiE/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lNDJFMDdmw +++ mktemp ++ local LAST_ERR=/tmp/tmp.mGEy8yqSOa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lNDJFMDdmw ++ cat /tmp/tmp.mGEy8yqSOa ++ rm /tmp/tmp.lNDJFMDdmw /tmp/tmp.mGEy8yqSOa ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/select-1.sql /tmp/tmp.U7OmJ8sNiE/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5tygzVqdhH +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZBgR6lOJjt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5tygzVqdhH ++ cat /tmp/tmp.ZBgR6lOJjt Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.5tygzVqdhH /tmp/tmp.ZBgR6lOJjt ++ return 0 + '[' '' ']' + desc 'check users created on cluster creation' + set +o xtrace ----------------------------------------------------------------------------------- check users created on cluster creation ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-haproxy -uuser-one -ptestpass' + local 'uri=-h some-name-haproxy -uuser-one -ptestpass' + local postfix= ++ echo -h some-name-haproxy -uuser-one -ptestpass ++ sed -e 's/.*-u//; s/ .*//' + local user=user-one + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-one -ptestpass' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-one -ptestpass' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8YRajpZM7H +++ mktemp ++ local LAST_ERR=/tmp/tmp.YdFuCcGSKa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8YRajpZM7H ++ cat /tmp/tmp.YdFuCcGSKa ++ rm /tmp/tmp.8YRajpZM7H /tmp/tmp.YdFuCcGSKa ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one.sql /tmp/tmp.U7OmJ8sNiE/user-one.sql + compare_mysql_cmd user-one-1 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1uGmX1pJA4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.hjHeluCwbb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1uGmX1pJA4 ++ cat /tmp/tmp.hjHeluCwbb ++ rm /tmp/tmp.1uGmX1pJA4 /tmp/tmp.hjHeluCwbb ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-one-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-1.sql /tmp/tmp.U7OmJ8sNiE/user-one-1.sql + compare_mysql_cmd user-one-2 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-2 + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eyHpaNo6LS +++ mktemp ++ local LAST_ERR=/tmp/tmp.lyS9hZd0el ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eyHpaNo6LS ++ cat /tmp/tmp.lyS9hZd0el ++ rm /tmp/tmp.eyHpaNo6LS /tmp/tmp.lyS9hZd0el ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-one-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-2.sql /tmp/tmp.U7OmJ8sNiE/user-one-2.sql + compare_mysql_user '-h some-name-haproxy -uuser-two -ptestpass3' + local 'uri=-h some-name-haproxy -uuser-two -ptestpass3' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-haproxy -uuser-two -ptestpass3 + local user=user-two + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-two -ptestpass3' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-two -ptestpass3' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ local LAST_OUT=/tmp/tmp.dRnfU5E4rP +++ mktemp ++ local LAST_ERR=/tmp/tmp.PGKAm8zANI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dRnfU5E4rP ++ cat /tmp/tmp.PGKAm8zANI ++ rm /tmp/tmp.dRnfU5E4rP /tmp/tmp.PGKAm8zANI ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two.sql /tmp/tmp.U7OmJ8sNiE/user-two.sql + compare_mysql_cmd user-two-1 'SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-two-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-two'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z0oEYxDmw6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.dZ5d8tTBp4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Z0oEYxDmw6 ++ cat /tmp/tmp.dZ5d8tTBp4 ++ rm /tmp/tmp.Z0oEYxDmw6 /tmp/tmp.dZ5d8tTBp4 ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-two-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two-1.sql /tmp/tmp.U7OmJ8sNiE/user-two-1.sql + generatedUserSecret=some-name-custom-user-secret ++ base64 -d ++ kubectl_bin get secret some-name-custom-user-secret -o 'jsonpath={.data.user-three}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UbzVjNETGU +++ mktemp ++ local LAST_ERR=/tmp/tmp.H0UwJq8tjb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secret some-name-custom-user-secret -o 'jsonpath={.data.user-three}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UbzVjNETGU ++ cat /tmp/tmp.H0UwJq8tjb ++ rm /tmp/tmp.UbzVjNETGU /tmp/tmp.H0UwJq8tjb ++ return 0 + userThreePass='Gi-9QFC{)Qr%ZzUaFBL' + compare_mysql_user '-h some-name-haproxy -uuser-three -p'\''Gi-9QFC{)Qr%ZzUaFBL'\''' + local 'uri=-h some-name-haproxy -uuser-three -p'\''Gi-9QFC{)Qr%ZzUaFBL'\''' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-haproxy -uuser-three '-p'\''Gi-9QFC{)Qr%ZzUaFBL'\''' + local user=user-three + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-three.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-three-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-three -p'\''Gi-9QFC{)Qr%ZzUaFBL'\''' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-three -p'\''Gi-9QFC{)Qr%ZzUaFBL'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PgOI9faUiW +++ mktemp ++ local LAST_ERR=/tmp/tmp.BccUeWHvhH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PgOI9faUiW ++ cat /tmp/tmp.BccUeWHvhH ++ rm /tmp/tmp.PgOI9faUiW /tmp/tmp.BccUeWHvhH ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-s5vdf + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-three.sql /tmp/tmp.U7OmJ8sNiE/user-three.sql + compare_mysql_cmd user-three-1 'SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-three-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-three-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-three-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-three'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1ReNHKIO4c +++ mktemp ++ local LAST_ERR=/tmp/tmp.zpqayQiJue ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1ReNHKIO4c ++ cat /tmp/tmp.zpqayQiJue ++ rm /tmp/tmp.1ReNHKIO4c /tmp/tmp.zpqayQiJue ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-three-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-three-1.sql /tmp/tmp.U7OmJ8sNiE/user-three-1.sql + desc 'check password change' + set +o xtrace ----------------------------------------------------------------------------------- check password change ----------------------------------------------------------------------------------- + kubectl_bin patch secret user-secrets '-p={"stringData":{"pwd-key-one": "new-password"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.gl15x7kAdX ++ mktemp + local LAST_ERR=/tmp/tmp.BMKWxIWicW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret user-secrets '-p={"stringData":{"pwd-key-one": "new-password"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.gl15x7kAdX secret/user-secrets patched + cat /tmp/tmp.BMKWxIWicW + rm /tmp/tmp.gl15x7kAdX /tmp/tmp.BMKWxIWicW + return 0 + sleep 15 + compare_mysql_user '-h some-name-haproxy -uuser-one -pnew-password' + local 'uri=-h some-name-haproxy -uuser-one -pnew-password' + local postfix= ++ echo -h some-name-haproxy -uuser-one -pnew-password ++ sed -e 's/.*-u//; s/ .*//' + local user=user-one + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-one -pnew-password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-one -pnew-password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ucM9Z7fBUf +++ mktemp ++ local LAST_ERR=/tmp/tmp.PxhyCOMinT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ucM9Z7fBUf ++ cat /tmp/tmp.PxhyCOMinT ++ rm /tmp/tmp.ucM9Z7fBUf /tmp/tmp.PxhyCOMinT ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one.sql /tmp/tmp.U7OmJ8sNiE/user-one.sql + compare_mysql_cmd user-one-1 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-one'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.99X2qpeqHn +++ mktemp ++ local LAST_ERR=/tmp/tmp.TG7zcwEY1A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.99X2qpeqHn ++ cat /tmp/tmp.TG7zcwEY1A ++ rm /tmp/tmp.99X2qpeqHn /tmp/tmp.TG7zcwEY1A ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-one-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-1.sql /tmp/tmp.U7OmJ8sNiE/user-one-1.sql + compare_mysql_cmd user-one-2 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-one-2 + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-one'\''@'\''127.0.0.1'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qHChbQKhKn +++ mktemp ++ local LAST_ERR=/tmp/tmp.izzaLeVIC2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qHChbQKhKn ++ cat /tmp/tmp.izzaLeVIC2 ++ rm /tmp/tmp.qHChbQKhKn /tmp/tmp.izzaLeVIC2 ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-s5vdf + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-one-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-2.sql /tmp/tmp.U7OmJ8sNiE/user-one-2.sql + desc 'check removing secretPasswordRef from user generates user password stored in generated custom-user-secret' + set +o xtrace ----------------------------------------------------------------------------------- check removing secretPasswordRef from user generates user password stored in generated custom-user-secret ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=json '-p=[{"op": "replace", "path": "/spec/users/1/passwordSecretRef", "value": null}]' ++ mktemp + local LAST_OUT=/tmp/tmp.6BN02HJ2sZ ++ mktemp + local LAST_ERR=/tmp/tmp.Y4EzEP5OZJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=json '-p=[{"op": "replace", "path": "/spec/users/1/passwordSecretRef", "value": null}]' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6BN02HJ2sZ perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.Y4EzEP5OZJ + rm /tmp/tmp.6BN02HJ2sZ /tmp/tmp.Y4EzEP5OZJ + return 0 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UUlAnUjYLP +++ mktemp ++ local LAST_ERR=/tmp/tmp.qOFxhnO1M0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UUlAnUjYLP ++ cat /tmp/tmp.qOFxhnO1M0 ++ rm /tmp/tmp.UUlAnUjYLP /tmp/tmp.qOFxhnO1M0 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GkVKn0TuuS +++ mktemp ++ local LAST_ERR=/tmp/tmp.qQmWGqlDJO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GkVKn0TuuS ++ cat /tmp/tmp.qQmWGqlDJO ++ rm /tmp/tmp.GkVKn0TuuS /tmp/tmp.qQmWGqlDJO ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.BBpWVinwp0 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.ytrMGABzkh +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.BBpWVinwp0 +++++ cat /tmp/tmp.ytrMGABzkh +++++ rm /tmp/tmp.BBpWVinwp0 /tmp/tmp.ytrMGABzkh +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r6KQsfI2iv +++ mktemp ++ local LAST_ERR=/tmp/tmp.hv4qOZd9VG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.r6KQsfI2iv ++ cat /tmp/tmp.hv4qOZd9VG ++ rm /tmp/tmp.r6KQsfI2iv /tmp/tmp.hv4qOZd9VG ++ return 0 + [[ 3 == \3 ]] + echo ++ base64 -d ++ kubectl_bin get secret some-name-custom-user-secret -o 'jsonpath={.data.user-two}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5UwDlTVhHj +++ mktemp ++ local LAST_ERR=/tmp/tmp.sHG851Lopf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secret some-name-custom-user-secret -o 'jsonpath={.data.user-two}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5UwDlTVhHj ++ cat /tmp/tmp.sHG851Lopf ++ rm /tmp/tmp.5UwDlTVhHj /tmp/tmp.sHG851Lopf ++ return 0 + userTwoPass='}#Fwf}aDR=rn45zI' + compare_mysql_user '-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' + local 'uri=-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' + local postfix= ++ echo -h some-name-haproxy -uuser-two '-p'\''}#Fwf}aDR=rn45zI'\''' ++ sed -e 's/.*-u//; s/ .*//' + local user=user-two + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.E9x4IR7ESm +++ mktemp ++ local LAST_ERR=/tmp/tmp.4jY9pVHJJO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.E9x4IR7ESm ++ cat /tmp/tmp.4jY9pVHJJO ++ rm /tmp/tmp.E9x4IR7ESm /tmp/tmp.4jY9pVHJJO ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-s5vdf + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two.sql /tmp/tmp.U7OmJ8sNiE/user-two.sql + desc 'delete initial users from CR and create a new one' + set +o xtrace ----------------------------------------------------------------------------------- delete initial users from CR and create a new one ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-four", "dbs": ["db1", "db2"], "grants":["SELECT, UPDATE"], "hosts": ["%"] } ]} }' ++ mktemp + local LAST_OUT=/tmp/tmp.LsQQfOecYX ++ mktemp + local LAST_ERR=/tmp/tmp.R9fLgsRORs + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-four", "dbs": ["db1", "db2"], "grants":["SELECT, UPDATE"], "hosts": ["%"] } ]} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LsQQfOecYX perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.R9fLgsRORs + rm /tmp/tmp.LsQQfOecYX /tmp/tmp.R9fLgsRORs + return 0 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SyAPNIGkBi +++ mktemp ++ local LAST_ERR=/tmp/tmp.2aGKi45S8x ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SyAPNIGkBi ++ cat /tmp/tmp.2aGKi45S8x ++ rm /tmp/tmp.SyAPNIGkBi /tmp/tmp.2aGKi45S8x ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PTZzHe8o9O +++ mktemp ++ local LAST_ERR=/tmp/tmp.PV4am34xhQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PTZzHe8o9O ++ cat /tmp/tmp.PV4am34xhQ ++ rm /tmp/tmp.PTZzHe8o9O /tmp/tmp.PV4am34xhQ ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Egs5iLrlc2 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.wBaW2HCkeW +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Egs5iLrlc2 +++++ cat /tmp/tmp.wBaW2HCkeW +++++ rm /tmp/tmp.Egs5iLrlc2 /tmp/tmp.wBaW2HCkeW +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VrDwDZVTnt +++ mktemp ++ local LAST_ERR=/tmp/tmp.YVbnuy7CWU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VrDwDZVTnt ++ cat /tmp/tmp.YVbnuy7CWU ++ rm /tmp/tmp.VrDwDZVTnt /tmp/tmp.YVbnuy7CWU ++ return 0 + [[ 3 == \3 ]] + echo ++ base64 -d ++ kubectl_bin get secret some-name-custom-user-secret -o 'jsonpath={.data.user-four}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OJYvBXsWXW +++ mktemp ++ local LAST_ERR=/tmp/tmp.9Y5Pi7QiLF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secret some-name-custom-user-secret -o 'jsonpath={.data.user-four}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OJYvBXsWXW ++ cat /tmp/tmp.9Y5Pi7QiLF ++ rm /tmp/tmp.OJYvBXsWXW /tmp/tmp.9Y5Pi7QiLF ++ return 0 + userFourPass='$yAz-2UJG2q$a>Q32Q}' + compare_mysql_user '-h some-name-haproxy -uuser-four -p'\''$yAz-2UJG2q$a>Q32Q}'\''' + local 'uri=-h some-name-haproxy -uuser-four -p'\''$yAz-2UJG2q$a>Q32Q}'\''' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-haproxy -uuser-four '-p'\''$yAz-2UJG2q$a>Q32Q}'\''' + local user=user-four + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-four -p'\''$yAz-2UJG2q$a>Q32Q}'\''' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-four -p'\''$yAz-2UJG2q$a>Q32Q}'\''' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NBnhejdwrR +++ mktemp ++ local LAST_ERR=/tmp/tmp.hLmNzuhAdt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NBnhejdwrR ++ cat /tmp/tmp.hLmNzuhAdt ++ rm /tmp/tmp.NBnhejdwrR /tmp/tmp.hLmNzuhAdt ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four.sql /tmp/tmp.U7OmJ8sNiE/user-four.sql + compare_mysql_cmd user-four-1 'SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-four-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jw1yTvwup9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PAQKHcupQm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jw1yTvwup9 ++ cat /tmp/tmp.PAQKHcupQm ++ rm /tmp/tmp.jw1yTvwup9 /tmp/tmp.PAQKHcupQm ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-four-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1.sql /tmp/tmp.U7OmJ8sNiE/user-four-1.sql + compare_mysql_user '-h some-name-haproxy -uuser-one -pnew-password' + local 'uri=-h some-name-haproxy -uuser-one -pnew-password' + local postfix= ++ echo -h some-name-haproxy -uuser-one -pnew-password ++ sed -e 's/.*-u//; s/ .*//' + local user=user-one + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-one -pnew-password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-one -pnew-password' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cwCbKiubCn +++ mktemp ++ local LAST_ERR=/tmp/tmp.yggk4v1xu8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cwCbKiubCn ++ cat /tmp/tmp.yggk4v1xu8 ++ rm /tmp/tmp.cwCbKiubCn /tmp/tmp.yggk4v1xu8 ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-one.sql /tmp/tmp.U7OmJ8sNiE/user-one.sql + compare_mysql_user '-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' + local 'uri=-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' + local postfix= ++ echo -h some-name-haproxy -uuser-two '-p'\''}#Fwf}aDR=rn45zI'\''' ++ sed -e 's/.*-u//; s/ .*//' + local user=user-two + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-two -p'\''}#Fwf}aDR=rn45zI'\''' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uaHmS278nt +++ mktemp ++ local LAST_ERR=/tmp/tmp.PwXzvGpKLu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uaHmS278nt ++ cat /tmp/tmp.PwXzvGpKLu ++ rm /tmp/tmp.uaHmS278nt /tmp/tmp.PwXzvGpKLu ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-two.sql /tmp/tmp.U7OmJ8sNiE/user-two.sql + desc 'check user DBs updated' + set +o xtrace ----------------------------------------------------------------------------------- check user DBs updated ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-four", "dbs": ["db1", "db2", "db3"], "grants":["SELECT, UPDATE"], "hosts": ["%"] } ]} }' ++ mktemp + local LAST_OUT=/tmp/tmp.cOGmCvyp0l ++ mktemp + local LAST_ERR=/tmp/tmp.8JvyL2SlZT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-four", "dbs": ["db1", "db2", "db3"], "grants":["SELECT, UPDATE"], "hosts": ["%"] } ]} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cOGmCvyp0l perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.8JvyL2SlZT + rm /tmp/tmp.cOGmCvyp0l /tmp/tmp.8JvyL2SlZT + return 0 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r4DS9JVAGH +++ mktemp ++ local LAST_ERR=/tmp/tmp.R6jO9zoPHv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.r4DS9JVAGH ++ cat /tmp/tmp.R6jO9zoPHv ++ rm /tmp/tmp.r4DS9JVAGH /tmp/tmp.R6jO9zoPHv ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y6ctrXZOla +++ mktemp ++ local LAST_ERR=/tmp/tmp.KOw4oMZxYB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y6ctrXZOla ++ cat /tmp/tmp.KOw4oMZxYB ++ rm /tmp/tmp.Y6ctrXZOla /tmp/tmp.KOw4oMZxYB ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.B57w6rmxe2 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.JNo62dwnoQ +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.B57w6rmxe2 +++++ cat /tmp/tmp.JNo62dwnoQ +++++ rm /tmp/tmp.B57w6rmxe2 /tmp/tmp.JNo62dwnoQ +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vT1hQacq8p +++ mktemp ++ local LAST_ERR=/tmp/tmp.WE7GfoSEZz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vT1hQacq8p ++ cat /tmp/tmp.WE7GfoSEZz ++ rm /tmp/tmp.vT1hQacq8p /tmp/tmp.WE7GfoSEZz ++ return 0 + [[ 3 == \3 ]] + echo + compare_mysql_cmd user-four-2 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-four-2 + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zgf8SgIL7H +++ mktemp ++ local LAST_ERR=/tmp/tmp.UiMfH9croB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zgf8SgIL7H ++ cat /tmp/tmp.UiMfH9croB ++ rm /tmp/tmp.zgf8SgIL7H /tmp/tmp.UiMfH9croB ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-four-2.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-2.sql /tmp/tmp.U7OmJ8sNiE/user-four-2.sql + desc 'check user grants updated' + set +o xtrace ----------------------------------------------------------------------------------- check user grants updated ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-four", "dbs": ["db1", "db2", "db3"], "grants":["SELECT, UPDATE, DELETE"], "hosts": ["%"] } ]} }' ++ mktemp + local LAST_OUT=/tmp/tmp.nWGAI2mb6i ++ mktemp + local LAST_ERR=/tmp/tmp.RVcKD7prJg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-four", "dbs": ["db1", "db2", "db3"], "grants":["SELECT, UPDATE, DELETE"], "hosts": ["%"] } ]} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nWGAI2mb6i perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.RVcKD7prJg + rm /tmp/tmp.nWGAI2mb6i /tmp/tmp.RVcKD7prJg + return 0 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XMHQ2RMLff +++ mktemp ++ local LAST_ERR=/tmp/tmp.MtevgeKoGn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XMHQ2RMLff ++ cat /tmp/tmp.MtevgeKoGn ++ rm /tmp/tmp.XMHQ2RMLff /tmp/tmp.MtevgeKoGn ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Vl3yrI5f5z +++ mktemp ++ local LAST_ERR=/tmp/tmp.cz3m6OfkiL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Vl3yrI5f5z ++ cat /tmp/tmp.cz3m6OfkiL ++ rm /tmp/tmp.Vl3yrI5f5z /tmp/tmp.cz3m6OfkiL ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.AJs5bGMm97 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Tml78Af6wJ +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.AJs5bGMm97 +++++ cat /tmp/tmp.Tml78Af6wJ +++++ rm /tmp/tmp.AJs5bGMm97 /tmp/tmp.Tml78Af6wJ +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GG9sHD9sSj +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lje8mCBAtr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GG9sHD9sSj ++ cat /tmp/tmp.Lje8mCBAtr ++ rm /tmp/tmp.GG9sHD9sSj /tmp/tmp.Lje8mCBAtr ++ return 0 + [[ 3 == \3 ]] + echo + compare_mysql_cmd user-four-3 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-four-3 + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A17zY5H55Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.EL7dErZwfA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.A17zY5H55Q ++ cat /tmp/tmp.EL7dErZwfA ++ rm /tmp/tmp.A17zY5H55Q /tmp/tmp.EL7dErZwfA ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-four-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-3.sql /tmp/tmp.U7OmJ8sNiE/user-four-3.sql + desc 'check user recreated after deleted from DB' + set +o xtrace ----------------------------------------------------------------------------------- check user recreated after deleted from DB ----------------------------------------------------------------------------------- + run_mysql 'DROP USER '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=DROP USER '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YkoBWwy6uY +++ mktemp ++ local LAST_ERR=/tmp/tmp.YPGIKHRTWz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YkoBWwy6uY ++ cat /tmp/tmp.YPGIKHRTWz ++ rm /tmp/tmp.YkoBWwy6uY /tmp/tmp.YPGIKHRTWz ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 15 + compare_mysql_cmd user-four-1 'SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-four-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YrZromBnlc +++ mktemp ++ local LAST_ERR=/tmp/tmp.S6Wbq0yv0W ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YrZromBnlc ++ cat /tmp/tmp.S6Wbq0yv0W ++ rm /tmp/tmp.YrZromBnlc /tmp/tmp.S6Wbq0yv0W ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-four-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1.sql /tmp/tmp.U7OmJ8sNiE/user-four-1.sql + compare_mysql_cmd user-four-3 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-four-3 + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SBtZetcd8N +++ mktemp ++ local LAST_ERR=/tmp/tmp.wFZEP34KG8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SBtZetcd8N ++ cat /tmp/tmp.wFZEP34KG8 ++ rm /tmp/tmp.SBtZetcd8N /tmp/tmp.wFZEP34KG8 ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-four-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-3.sql /tmp/tmp.U7OmJ8sNiE/user-four-3.sql + desc 'check user update from DB' + set +o xtrace ----------------------------------------------------------------------------------- check user update from DB ----------------------------------------------------------------------------------- + run_mysql 'GRANT INSERT ON db1.* TO '\''user-four'\''@'\''%'\''' '-h some-name-haproxy -uroot -proot_password' + local 'command=GRANT INSERT ON db1.* TO '\''user-four'\''@'\''%'\''' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dyL0PXoUPJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.W83KpIHZYE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dyL0PXoUPJ ++ cat /tmp/tmp.W83KpIHZYE ++ rm /tmp/tmp.dyL0PXoUPJ /tmp/tmp.W83KpIHZYE ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 15 + compare_mysql_cmd user-four-1 'SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-four-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-four'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z0IsXHyJ9P +++ mktemp ++ local LAST_ERR=/tmp/tmp.plmbttq9ze ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Z0IsXHyJ9P ++ cat /tmp/tmp.plmbttq9ze ++ rm /tmp/tmp.Z0IsXHyJ9P /tmp/tmp.plmbttq9ze ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-four-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-1.sql /tmp/tmp.U7OmJ8sNiE/user-four-1.sql + compare_mysql_cmd user-four-4 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-four-4 + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-4-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SHOW GRANTS FOR '\''user-four'\''@'\''%'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uFsDdXO8nJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.TvMJW9rEXJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uFsDdXO8nJ ++ cat /tmp/tmp.TvMJW9rEXJ ++ rm /tmp/tmp.uFsDdXO8nJ /tmp/tmp.TvMJW9rEXJ ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-four-4.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-four-4.sql /tmp/tmp.U7OmJ8sNiE/user-four-4.sql + desc 'check new user created after updated user name via CR' + set +o xtrace ----------------------------------------------------------------------------------- check new user created after updated user name via CR ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-five", "dbs": ["db1", "db2", "db3"], "grants":["SELECT, UPDATE, DELETE"], "hosts": ["%"] } ]} }' ++ mktemp + local LAST_OUT=/tmp/tmp.aLRvazvCUQ ++ mktemp + local LAST_ERR=/tmp/tmp.lHtrZfSqnm + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={ "spec": {"users":[ { "name":"user-five", "dbs": ["db1", "db2", "db3"], "grants":["SELECT, UPDATE, DELETE"], "hosts": ["%"] } ]} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.aLRvazvCUQ perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.lHtrZfSqnm + rm /tmp/tmp.aLRvazvCUQ /tmp/tmp.lHtrZfSqnm + return 0 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TR4q7bSALw +++ mktemp ++ local LAST_ERR=/tmp/tmp.bHunOerEP0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TR4q7bSALw ++ cat /tmp/tmp.bHunOerEP0 ++ rm /tmp/tmp.TR4q7bSALw /tmp/tmp.bHunOerEP0 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1KJWPEpNmS +++ mktemp ++ local LAST_ERR=/tmp/tmp.cE6Gr3Mf5V ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1KJWPEpNmS ++ cat /tmp/tmp.cE6Gr3Mf5V ++ rm /tmp/tmp.1KJWPEpNmS /tmp/tmp.cE6Gr3Mf5V ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.CVAQshsdZ6 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.aLoKghsZAO +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.CVAQshsdZ6 +++++ cat /tmp/tmp.aLoKghsZAO +++++ rm /tmp/tmp.CVAQshsdZ6 /tmp/tmp.aLoKghsZAO +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1zPEYobhkf +++ mktemp ++ local LAST_ERR=/tmp/tmp.piVm63b6bJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1zPEYobhkf ++ cat /tmp/tmp.piVm63b6bJ ++ rm /tmp/tmp.1zPEYobhkf /tmp/tmp.piVm63b6bJ ++ return 0 + [[ 3 == \3 ]] + echo ++ kubectl_bin get secret some-name-custom-user-secret -o 'jsonpath={.data.user-five}' ++ base64 -d +++ mktemp ++ local LAST_OUT=/tmp/tmp.ckEJTzdxPI +++ mktemp ++ local LAST_ERR=/tmp/tmp.SOnFejzfbu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secret some-name-custom-user-secret -o 'jsonpath={.data.user-five}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ckEJTzdxPI ++ cat /tmp/tmp.SOnFejzfbu ++ rm /tmp/tmp.ckEJTzdxPI /tmp/tmp.SOnFejzfbu ++ return 0 + userFivePass='=%9I<4E79zF1#O~[W' + compare_mysql_user '-h some-name-haproxy -uuser-five -p'\''=%9I<4E79zF1#O~[W'\''' + local 'uri=-h some-name-haproxy -uuser-five -p'\''=%9I<4E79zF1#O~[W'\''' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-haproxy -uuser-five '-p'\''=%9I<4E79zF1#O~[W'\''' + local user=user-five + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-five.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-five-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-haproxy -uuser-five -p'\''=%9I<4E79zF1#O~[W'\''' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-haproxy -uuser-five -p'\''=%9I<4E79zF1#O~[W'\''' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.92MmLufD55 +++ mktemp ++ local LAST_ERR=/tmp/tmp.sevvC2Al6m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.92MmLufD55 ++ cat /tmp/tmp.sevvC2Al6m ++ rm /tmp/tmp.92MmLufD55 /tmp/tmp.sevvC2Al6m ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-five.sql /tmp/tmp.U7OmJ8sNiE/user-five.sql + compare_mysql_cmd user-five-1 'SELECT User, Host from mysql.user WHERE User = '\''user-five'\'';' '-h some-name-haproxy -uroot -proot_password' + local command_id=user-five-1 + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-five'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-five-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-five-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT User, Host from mysql.user WHERE User = '\''user-five'\'';' '-h some-name-haproxy -uroot -proot_password' + local 'command=SELECT User, Host from mysql.user WHERE User = '\''user-five'\'';' + local 'uri=-h some-name-haproxy -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mOQAxklpxI +++ mktemp ++ local LAST_ERR=/tmp/tmp.jIKDY4kLEp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mOQAxklpxI ++ cat /tmp/tmp.jIKDY4kLEp ++ rm /tmp/tmp.mOQAxklpxI /tmp/tmp.jIKDY4kLEp ++ return 0 + client_pod=pxc-client-59944c5bbf-s5vdf + wait_pod pxc-client-59944c5bbf-s5vdf + local pod=pxc-client-59944c5bbf-s5vdf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-s5vdf ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-s5vdf condition met waiting for pod/pxc-client-59944c5bbf-s5vdf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.U7OmJ8sNiE/user-five-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/custom-users/compare/user-five-1.sql /tmp/tmp.U7OmJ8sNiE/user-five-1.sql + destroy custom-users-20369 + local namespace=custom-users-20369 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + sort -u + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + grep -v level=info + tee /tmp/tmp.U7OmJ8sNiE/operator.log ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator + grep -v 'the object has been modified' + grep -v 'get backup status: Job.batch' ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.V3iixn4r7E +++ mktemp ++ local LAST_ERR=/tmp/tmp.Cd5CvvlUd2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.V3iixn4r7E ++ cat /tmp/tmp.Cd5CvvlUd2 ++ rm /tmp/tmp.V3iixn4r7E /tmp/tmp.Cd5CvvlUd2 ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-58dd9fd94c-c49gt ++ mktemp + local LAST_OUT=/tmp/tmp.VZk2pRvAkl ++ mktemp + local LAST_ERR=/tmp/tmp.xZ967ffKMS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-58dd9fd94c-c49gt + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VZk2pRvAkl + cat /tmp/tmp.xZ967ffKMS + rm /tmp/tmp.VZk2pRvAkl /tmp/tmp.xZ967ffKMS + return 0 2025-11-06T12:16:32.748Z INFO setup Manager starting up {"gitCommit": "89209ce179be0afd4246cdcc6b564d43f706c45f", "gitBranch": "PR-2207-89209ce1", "buildTime": "2025-11-06T11:54:31Z", "goVersion": "go1.25.4", "os": "linux", "arch": "amd64"} 2025-11-06T12:16:32.748Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.31.13-gke.1139000"} 2025-11-06T12:16:32.752Z INFO setup Registering Components. 2025-11-06T12:16:33.945Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2025-11-06T12:16:33.945Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2025-11-06T12:16:33.945Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2025-11-06T12:16:33.945Z INFO controller-runtime.metrics Starting metrics server 2025-11-06T12:16:33.945Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2025-11-06T12:16:33.945Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2025-11-06T12:16:33.945Z INFO controller-runtime.webhook Starting webhook server 2025-11-06T12:16:33.945Z INFO setup Starting the Cmd. 2025-11-06T12:16:33.945Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2025-11-06T12:16:34.046Z INFO attempting to acquire leader lease pxc-operator/08db1feb.percona.com... 2025-11-06T12:16:34.080Z INFO successfully acquired lease pxc-operator/08db1feb.percona.com 2025-11-06T12:16:34.081Z DEBUG events percona-xtradb-cluster-operator-58dd9fd94c-c49gt_20016949-759b-40a9-b2cc-76f6474624c2 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"7b0c8023-516d-4419-b10b-a9897179abf6","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1762431394074383009"}, "reason": "LeaderElection"} 2025-11-06T12:16:34.081Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.PerconaXtraDBCluster"} 2025-11-06T12:16:34.081Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.Secret"} 2025-11-06T12:16:34.081Z INFO Starting EventSource {"controller": "pxcrestore-controller", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2025-11-06T12:16:34.082Z INFO Starting EventSource {"controller": "pxcbackup-controller", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2025-11-06T12:16:34.181Z INFO Starting Controller {"controller": "pxcrestore-controller"} 2025-11-06T12:16:34.181Z INFO Starting workers {"controller": "pxcrestore-controller", "worker count": 1} 2025-11-06T12:16:34.183Z INFO Starting Controller {"controller": "pxcbackup-controller"} 2025-11-06T12:16:34.183Z INFO Starting Controller {"controller": "pxc-controller"} 2025-11-06T12:16:34.183Z INFO Starting workers {"controller": "pxcbackup-controller", "worker count": 1} 2025-11-06T12:16:34.183Z INFO Starting workers {"controller": "pxc-controller", "worker count": 1} 2025-11-06T12:17:37.143Z INFO Set CR version {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "version": "1.19.0"} 2025-11-06T12:17:37.471Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-06T12:17:37.590Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-06T12:17:37.622Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-06T12:17:37.670Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-06T12:17:37.700Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-06T12:17:37.743Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-06T12:17:37.836Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "7dfa45a4-0db2-4aa7-b1f9-712b71e08aa9", "object": "some-name-haproxy-replicas", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-06T12:17:38.374Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "c6a82788-4449-4d34-b19f-231d031d11e5", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-06T12:17:38.402Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "c6a82788-4449-4d34-b19f-231d031d11e5", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-06T12:18:54.937Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "6e835bb4-a301-4b43-ac55-99a769207ed6", "user": "operator"} 2025-11-06T12:18:54.976Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "6e835bb4-a301-4b43-ac55-99a769207ed6", "user": "monitor"} 2025-11-06T12:18:55.039Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "6e835bb4-a301-4b43-ac55-99a769207ed6"} 2025-11-06T12:18:55.121Z INFO monitor user privileges granted {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "6e835bb4-a301-4b43-ac55-99a769207ed6"} 2025-11-06T12:18:55.167Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "6e835bb4-a301-4b43-ac55-99a769207ed6", "user": "xtrabackup"} 2025-11-06T12:18:55.222Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "6e835bb4-a301-4b43-ac55-99a769207ed6"} 2025-11-06T12:18:55.258Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "6e835bb4-a301-4b43-ac55-99a769207ed6", "user": "replication"} 2025-11-06T12:21:21.411Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "root"} 2025-11-06T12:21:21.453Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-one"} 2025-11-06T12:21:21.453Z INFO User not created {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-one"} 2025-11-06T12:21:21.519Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-two"} 2025-11-06T12:21:21.519Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-one"} 2025-11-06T12:21:21.519Z INFO User not created {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-two"} 2025-11-06T12:21:21.547Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-two"} 2025-11-06T12:21:21.565Z INFO Created custom user secrets {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "secrets": "my-cluster-secrets"} 2025-11-06T12:21:21.566Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-three"} 2025-11-06T12:21:21.566Z INFO User not created {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-three"} 2025-11-06T12:21:21.589Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "user": "user-three"} 2025-11-06T12:21:21.667Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0bf54b4b-ef00-4621-90a5-273181a8ef6f", "new version": "8.0.43-34.1"} 2025-11-06T12:23:51.530Z INFO User password changed {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "004c9e34-1063-4691-aba6-ce72ab7e6cea", "user": "user-one"} 2025-11-06T12:23:51.568Z INFO User password updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "004c9e34-1063-4691-aba6-ce72ab7e6cea", "user": "user-one"} 2025-11-06T12:24:28.017Z INFO User password changed {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "1bd22ea0-40af-45ac-b157-35b6f3a70e6e", "user": "user-two"} 2025-11-06T12:24:28.049Z INFO User password updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "1bd22ea0-40af-45ac-b157-35b6f3a70e6e", "user": "user-two"} 2025-11-06T12:24:55.585Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "fe7366a6-6e08-48ce-a7e0-4eae08cb5d4a", "user": "user-four"} 2025-11-06T12:24:55.585Z INFO User not created {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "fe7366a6-6e08-48ce-a7e0-4eae08cb5d4a", "user": "user-four"} 2025-11-06T12:24:55.632Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "fe7366a6-6e08-48ce-a7e0-4eae08cb5d4a", "user": "user-four"} 2025-11-06T12:25:51.351Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "24ca3ed9-2601-490a-a1a0-21439d62119c", "user": "user-four"} 2025-11-06T12:25:51.351Z INFO DBs changed {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "24ca3ed9-2601-490a-a1a0-21439d62119c", "current": {"db1":{},"db2":{}}, "desired": ["db1", "db2", "db3"], "user": "user-four"} 2025-11-06T12:25:51.414Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "24ca3ed9-2601-490a-a1a0-21439d62119c", "user": "user-four"} 2025-11-06T12:26:19.817Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "cfea8672-501a-4bd5-a4f7-1b8a77b3f13e", "user": "user-four"} 2025-11-06T12:26:19.817Z INFO Grant not present in current grants {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "cfea8672-501a-4bd5-a4f7-1b8a77b3f13e", "grant": "SELECT, UPDATE, DELETE", "user": "user-four"} 2025-11-06T12:26:19.870Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "cfea8672-501a-4bd5-a4f7-1b8a77b3f13e", "user": "user-four"} 2025-11-06T12:26:29.057Z INFO controller-runtime.cache Warning: watch ended with error {"reflector": "pkg/mod/k8s.io/client-go@v0.34.1/tools/cache/reflector.go:290", "type": "*v1.Event", "err": "very short watch: pkg/mod/k8s.io/client-go@v0.34.1/tools/cache/reflector.go:290: Unexpected watch close - watch lasted less than a second and no items received"} 2025-11-06T12:26:54.820Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "48e8dcc1-bed8-47d5-9c87-c54c11b977e5", "user": "user-four"} 2025-11-06T12:26:54.820Z INFO User not created {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "48e8dcc1-bed8-47d5-9c87-c54c11b977e5", "user": "user-four"} 2025-11-06T12:26:54.883Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "48e8dcc1-bed8-47d5-9c87-c54c11b977e5", "user": "user-four"} 2025-11-06T12:27:32.669Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "5c3807e2-6351-4478-9ed0-a22be7bd44cf", "user": "user-four"} 2025-11-06T12:27:32.669Z INFO Grant not present in current grants {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "5c3807e2-6351-4478-9ed0-a22be7bd44cf", "grant": "SELECT, UPDATE, DELETE", "user": "user-four"} 2025-11-06T12:27:32.739Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "5c3807e2-6351-4478-9ed0-a22be7bd44cf", "user": "user-four"} 2025-11-06T12:27:38.026Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "d6a32acd-dbb8-4e0e-ae24-23848ad3ba26", "user": "user-four"} 2025-11-06T12:27:38.026Z INFO Grant not present in current grants {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "d6a32acd-dbb8-4e0e-ae24-23848ad3ba26", "grant": "SELECT, UPDATE, DELETE", "user": "user-four"} 2025-11-06T12:27:38.097Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "d6a32acd-dbb8-4e0e-ae24-23848ad3ba26", "user": "user-four"} 2025-11-06T12:27:43.370Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "8ad926b6-d2a9-4fa6-9bcf-2093c03a30f1", "user": "user-four"} 2025-11-06T12:27:43.370Z INFO Grant not present in current grants {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "8ad926b6-d2a9-4fa6-9bcf-2093c03a30f1", "grant": "SELECT, UPDATE, DELETE", "user": "user-four"} 2025-11-06T12:27:43.442Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "8ad926b6-d2a9-4fa6-9bcf-2093c03a30f1", "user": "user-four"} 2025-11-06T12:27:48.768Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "e40ce1df-a37d-4ec9-a6e2-de344026e7c2", "user": "user-four"} 2025-11-06T12:27:48.768Z INFO Grant not present in current grants {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "e40ce1df-a37d-4ec9-a6e2-de344026e7c2", "grant": "SELECT, UPDATE, DELETE", "user": "user-four"} 2025-11-06T12:27:48.843Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "e40ce1df-a37d-4ec9-a6e2-de344026e7c2", "user": "user-four"} 2025-11-06T12:27:54.138Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "194f2c23-2060-4ed6-b0a6-140e77b1f91c", "user": "user-four"} 2025-11-06T12:27:54.138Z INFO Grant not present in current grants {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "194f2c23-2060-4ed6-b0a6-140e77b1f91c", "grant": "SELECT, UPDATE, DELETE", "user": "user-four"} 2025-11-06T12:27:54.207Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "194f2c23-2060-4ed6-b0a6-140e77b1f91c", "user": "user-four"} 2025-11-06T12:27:59.497Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0ea90a9b-5755-472c-80a7-d6d385242c8e", "user": "user-four"} 2025-11-06T12:27:59.497Z INFO Grant not present in current grants {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0ea90a9b-5755-472c-80a7-d6d385242c8e", "grant": "SELECT, UPDATE, DELETE", "user": "user-four"} 2025-11-06T12:27:59.567Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "0ea90a9b-5755-472c-80a7-d6d385242c8e", "user": "user-four"} 2025-11-06T12:28:01.124Z INFO Creating/updating user {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "82bcbe64-fa40-4a6b-9c9d-595cb39948a0", "user": "user-five"} 2025-11-06T12:28:01.124Z INFO User not created {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "82bcbe64-fa40-4a6b-9c9d-595cb39948a0", "user": "user-five"} 2025-11-06T12:28:01.186Z INFO User created/updated {"controller": "pxc-controller", "namespace": "custom-users-20369", "name": "some-name", "reconcileID": "82bcbe64-fa40-4a6b-9c9d-595cb39948a0", "user": "user-five"} + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch pxc -n custom-users-20369 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.2dHAxUQMwm ++ mktemp + local LAST_ERR=/tmp/tmp.SzJfcwcXfO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2dHAxUQMwm perconaxtradbcluster.pxc.percona.com "some-name" deleted from custom-users-20369 namespace + cat /tmp/tmp.SzJfcwcXfO + rm /tmp/tmp.2dHAxUQMwm /tmp/tmp.SzJfcwcXfO + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.Uy7gYj7Xpu ++ mktemp + local LAST_ERR=/tmp/tmp.StbK7FNhsH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Uy7gYj7Xpu No resources found + cat /tmp/tmp.StbK7FNhsH + rm /tmp/tmp.Uy7gYj7Xpu /tmp/tmp.StbK7FNhsH + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.uYthDtu8Jv ++ mktemp + local LAST_ERR=/tmp/tmp.wCOiTdLq5T + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uYthDtu8Jv No resources found + cat /tmp/tmp.wCOiTdLq5T + rm /tmp/tmp.uYthDtu8Jv /tmp/tmp.wCOiTdLq5T + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.2BdUV6RWkj ++ mktemp + local LAST_ERR=/tmp/tmp.8vRMd7VdDV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2BdUV6RWkj validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.8vRMd7VdDV + rm /tmp/tmp.2BdUV6RWkj /tmp/tmp.8vRMd7VdDV + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace custom-users-20369 + rm -rf /tmp/tmp.U7OmJ8sNiE ++ mktemp + local LAST_OUT=/tmp/tmp.uh8RGqNtiH ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_ERR=/tmp/tmp.jDblzHCqOM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace custom-users-20369 + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.i02Oe75Pxr ++ mktemp + local LAST_ERR=/tmp/tmp.wBFr5eTXGr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator