Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/logs/users-5-7.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + create_infra users-3672 + local ns=users-3672 + '[' -n pxc-operator ']' + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + kubectl patch pxc -n users-32560 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.QNMSncqZdy ++ mktemp + local LAST_ERR=/tmp/tmp.azE8qf9o74 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QNMSncqZdy perconaxtradbcluster.pxc.percona.com "some-name" deleted from users-32560 namespace + cat /tmp/tmp.azE8qf9o74 + rm /tmp/tmp.QNMSncqZdy /tmp/tmp.azE8qf9o74 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.VKfYhIALtK ++ mktemp + local LAST_ERR=/tmp/tmp.zkqSd3vPs8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VKfYhIALtK No resources found + cat /tmp/tmp.zkqSd3vPs8 + rm /tmp/tmp.VKfYhIALtK /tmp/tmp.zkqSd3vPs8 + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.VHFI7xlblT ++ mktemp + local LAST_ERR=/tmp/tmp.M4zurJ6xck + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VHFI7xlblT No resources found + cat /tmp/tmp.M4zurJ6xck + rm /tmp/tmp.VHFI7xlblT /tmp/tmp.M4zurJ6xck + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.srB7ua6vo9 + awk '{print$1}' + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.BHy5vmyMhB ++ mktemp + local LAST_ERR=/tmp/tmp.utGzHaD7eb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns ++ mktemp + local LAST_ERR=/tmp/tmp.bg7XD9TzFc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + xargs kubectl delete ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BHy5vmyMhB + cat /tmp/tmp.utGzHaD7eb + rm /tmp/tmp.BHy5vmyMhB /tmp/tmp.utGzHaD7eb + return 0 namespace "users-32560" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.srB7ua6vo9 namespace "pxc-operator" deleted + cat /tmp/tmp.bg7XD9TzFc + rm /tmp/tmp.srB7ua6vo9 /tmp/tmp.bg7XD9TzFc + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.nfKpSVtpkr ++ mktemp + local LAST_ERR=/tmp/tmp.eSiDRVs83j + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nfKpSVtpkr namespace/pxc-operator created + cat /tmp/tmp.eSiDRVs83j + rm /tmp/tmp.nfKpSVtpkr /tmp/tmp.eSiDRVs83j + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.jaV5STKwuB +++ mktemp ++ local LAST_ERR=/tmp/tmp.v5xRmJAQ4v ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jaV5STKwuB ++ cat /tmp/tmp.v5xRmJAQ4v ++ rm /tmp/tmp.jaV5STKwuB /tmp/tmp.v5xRmJAQ4v ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2232-31b0cef6-17-cluster6 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.jo35ohOfup ++ mktemp + local LAST_ERR=/tmp/tmp.RJHF7DIV8I + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2232-31b0cef6-17-cluster6 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jo35ohOfup Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2232-31b0cef6-17-cluster6" modified. + cat /tmp/tmp.RJHF7DIV8I + rm /tmp/tmp.jo35ohOfup /tmp/tmp.RJHF7DIV8I + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.dZeKYma6og ++ mktemp + local LAST_ERR=/tmp/tmp.VJ5gFfwMLe + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dZeKYma6og customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.VJ5gFfwMLe + rm /tmp/tmp.dZeKYma6og /tmp/tmp.VJ5gFfwMLe + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.T3Q62EGF4n ++ mktemp + local LAST_ERR=/tmp/tmp.vWNvcmRzcz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.T3Q62EGF4n clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.vWNvcmRzcz + rm /tmp/tmp.T3Q62EGF4n /tmp/tmp.vWNvcmRzcz + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2232-31b0cef6^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + kubectl_bin apply -f - ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/deploy/cw-operator.yaml + local LAST_OUT=/tmp/tmp.7ADiIVYR37 ++ mktemp + local LAST_ERR=/tmp/tmp.EFImIn58XN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.7ADiIVYR37 deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.EFImIn58XN + rm /tmp/tmp.7ADiIVYR37 /tmp/tmp.EFImIn58XN + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.750aq24tei ++ mktemp + local LAST_ERR=/tmp/tmp.KHtGDj0p0K + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.750aq24tei pod/percona-xtradb-cluster-operator-58ffd6cd64-hgk7p condition met + cat /tmp/tmp.KHtGDj0p0K + rm /tmp/tmp.750aq24tei /tmp/tmp.KHtGDj0p0K + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.hRcB8sZNAC +++ mktemp ++ local LAST_ERR=/tmp/tmp.4ErhTY4JTc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hRcB8sZNAC ++ cat /tmp/tmp.4ErhTY4JTc ++ rm /tmp/tmp.hRcB8sZNAC /tmp/tmp.4ErhTY4JTc ++ return 0 + wait_pod percona-xtradb-cluster-operator-58ffd6cd64-hgk7p 480 pxc-operator + local pod=percona-xtradb-cluster-operator-58ffd6cd64-hgk7p + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-58ffd6cd64-hgk7p ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-58ffd6cd64-hgk7p condition met waiting for pod/percona-xtradb-cluster-operator-58ffd6cd64-hgk7p to become Ready.Ok + sleep 3 + create_namespace users-3672 + local namespace=users-3672 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces users-3672' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces users-3672 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace users-3672 ++ mktemp ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.xbcenmxdK9 + local LAST_OUT=/tmp/tmp.LbTGMVaL35 + xargs kubectl delete ns ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.IsTMiECc2M + local exit_status=0 + local LAST_ERR=/tmp/tmp.SUnQRk3GMO + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-3672 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xbcenmxdK9 + cat /tmp/tmp.IsTMiECc2M + rm /tmp/tmp.xbcenmxdK9 /tmp/tmp.IsTMiECc2M + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-3672 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-3672 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.LbTGMVaL35 + cat /tmp/tmp.SUnQRk3GMO Error from server (NotFound): namespaces "users-3672" not found + rm /tmp/tmp.LbTGMVaL35 /tmp/tmp.SUnQRk3GMO + return 1 + : + wait_for_delete namespace/users-3672 + local res=namespace/users-3672 + echo -n 'waiting for namespace/users-3672 to be deleted' waiting for namespace/users-3672 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "users-3672" not found + desc 'create namespace users-3672' + set +o xtrace ----------------------------------------------------------------------------------- create namespace users-3672 ----------------------------------------------------------------------------------- + kubectl_bin create namespace users-3672 ++ mktemp + local LAST_OUT=/tmp/tmp.0LgEFxaIIF ++ mktemp + local LAST_ERR=/tmp/tmp.stEInHSrXD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace users-3672 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0LgEFxaIIF namespace/users-3672 created + cat /tmp/tmp.stEInHSrXD + rm /tmp/tmp.0LgEFxaIIF /tmp/tmp.stEInHSrXD + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z6EnQQXWKY +++ mktemp ++ local LAST_ERR=/tmp/tmp.13aIDCKX3A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Z6EnQQXWKY ++ cat /tmp/tmp.13aIDCKX3A ++ rm /tmp/tmp.Z6EnQQXWKY /tmp/tmp.13aIDCKX3A ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2232-31b0cef6-17-cluster6 --namespace=users-3672 ++ mktemp + local LAST_OUT=/tmp/tmp.VjDgZuqpgn ++ mktemp + local LAST_ERR=/tmp/tmp.oHP8FYwcia + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2232-31b0cef6-17-cluster6 --namespace=users-3672 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VjDgZuqpgn Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2232-31b0cef6-17-cluster6" modified. + cat /tmp/tmp.oHP8FYwcia + rm /tmp/tmp.VjDgZuqpgn /tmp/tmp.oHP8FYwcia + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.JIKCnpYVJN ++ mktemp + local LAST_ERR=/tmp/tmp.wbUmktwgvU + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JIKCnpYVJN secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.wbUmktwgvU + rm /tmp/tmp.JIKCnpYVJN /tmp/tmp.wbUmktwgvU + return 0 + desc 'create PXC cluster with 1-password secret' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster with 1-password secret ----------------------------------------------------------------------------------- + newpass=test-password ++ echo -n test-password ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZA== + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/some-name.yml '' '' /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/secrets_one_pass.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/secrets_one_pass.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/secrets_one_pass.yml ++ mktemp + local LAST_OUT=/tmp/tmp.p1Zypyxutl ++ mktemp + local LAST_ERR=/tmp/tmp.Hsm1jfA8Op + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/secrets_one_pass.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.p1Zypyxutl secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.Hsm1jfA8Op + rm /tmp/tmp.p1Zypyxutl /tmp/tmp.Hsm1jfA8Op + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/client.yml + kubectl_bin apply -f - + local pvc_name= + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/client.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + local LAST_OUT=/tmp/tmp.PxjJ7dRBFh + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' ++ mktemp + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2232-31b0cef6#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-3672~ + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_ERR=/tmp/tmp.QPDu1lWXAj + local exit_status=0 ++ seq 0 2 + /usr/bin/sed -e 's#apply:.*#apply: Never#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PxjJ7dRBFh deployment.apps/pxc-client created + cat /tmp/tmp.QPDu1lWXAj + rm /tmp/tmp.PxjJ7dRBFh /tmp/tmp.QPDu1lWXAj + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p\-\c\l\o\u\d ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/some-name.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/some-name.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/some-name.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/some-name.yml + kubectl_bin apply -f - + local pvc_name= + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + local LAST_OUT=/tmp/tmp.VdOjvoirxY + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' ++ mktemp + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-3672~ + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2232-31b0cef6#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + local LAST_ERR=/tmp/tmp.fcsIfBkiEn + local exit_status=0 + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VdOjvoirxY perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.fcsIfBkiEn + rm /tmp/tmp.VdOjvoirxY /tmp/tmp.fcsIfBkiEn + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6OgcrSXkzr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vmWRUJUkZI +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.6OgcrSXkzr +++ cat /tmp/tmp.vmWRUJUkZI +++ rm /tmp/tmp.6OgcrSXkzr /tmp/tmp.vmWRUJUkZI +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VK53DgT6DJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VQgAGliitF +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.VK53DgT6DJ +++ cat /tmp/tmp.VQgAGliitF +++ rm /tmp/tmp.VK53DgT6DJ /tmp/tmp.VQgAGliitF +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-3672 ++ mktemp + local LAST_OUT=/tmp/tmp.7cc5JvpWmI ++ mktemp + local LAST_ERR=/tmp/tmp.BJrf0JBgBc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-3672 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-3672 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-3672 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.7cc5JvpWmI + cat /tmp/tmp.BJrf0JBgBc error: no matching resources found + rm /tmp/tmp.7cc5JvpWmI /tmp/tmp.BJrf0JBgBc + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.CFY3jbjQg1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ElvouNscgn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CFY3jbjQg1 ++ cat /tmp/tmp.ElvouNscgn ++ rm /tmp/tmp.CFY3jbjQg1 /tmp/tmp.ElvouNscgn ++ return 0 + local 'root_pass=7OVnIPdog&]S>X-4' + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' +++ mktemp ++ grep -E -o 'early-plugin-load=keyring_\w+.so' ++ local LAST_OUT=/tmp/tmp.hr0IaXl0hT +++ mktemp ++ local LAST_ERR=/tmp/tmp.R4sm0TWIyM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hr0IaXl0hT ++ cat /tmp/tmp.R4sm0TWIyM Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.hr0IaXl0hT /tmp/tmp.R4sm0TWIyM ++ return 0 + [[ -n '' ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KtmjOEGVgh +++ mktemp ++ local LAST_ERR=/tmp/tmp.P1ueIn3JKM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KtmjOEGVgh ++ cat /tmp/tmp.P1ueIn3JKM ++ rm /tmp/tmp.KtmjOEGVgh /tmp/tmp.P1ueIn3JKM ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.toeHNSvbLZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.5tPg8vwqJx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.toeHNSvbLZ ++ cat /tmp/tmp.5tPg8vwqJx ++ rm /tmp/tmp.toeHNSvbLZ /tmp/tmp.5tPg8vwqJx ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5CGDpddpwN +++ mktemp ++ local LAST_ERR=/tmp/tmp.hzeBNpF47u ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5CGDpddpwN ++ cat /tmp/tmp.hzeBNpF47u ++ rm /tmp/tmp.5CGDpddpwN /tmp/tmp.hzeBNpF47u ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1.sql /tmp/tmp.miV4rrH5zD/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Mkf3Okl5cU +++ mktemp ++ local LAST_ERR=/tmp/tmp.cgIarGaaMi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Mkf3Okl5cU ++ cat /tmp/tmp.cgIarGaaMi ++ rm /tmp/tmp.Mkf3Okl5cU /tmp/tmp.cgIarGaaMi ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1.sql /tmp/tmp.miV4rrH5zD/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''7OVnIPdog&]S>X-4'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7gm6jRQrme +++ mktemp ++ local LAST_ERR=/tmp/tmp.ttCG12Lyl9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7gm6jRQrme ++ cat /tmp/tmp.ttCG12Lyl9 ++ rm /tmp/tmp.7gm6jRQrme /tmp/tmp.ttCG12Lyl9 ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-1.sql /tmp/tmp.miV4rrH5zD/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ grep -E -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MHAs6oUXkN +++ mktemp ++ local LAST_ERR=/tmp/tmp.txXApQLdKM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MHAs6oUXkN ++ cat /tmp/tmp.txXApQLdKM Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.MHAs6oUXkN /tmp/tmp.txXApQLdKM ++ return 0 + '[' '' ']' + desc 'test missing passwords were created and present in internal secrets' + set +o xtrace ----------------------------------------------------------------------------------- test missing passwords were created and present in internal secrets ----------------------------------------------------------------------------------- + empty_pwds=() + wrong_pwds=() + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking root' Checking root ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.3wi7p5EG7v +++ mktemp ++ local LAST_ERR=/tmp/tmp.lzS4XVzmSR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3wi7p5EG7v ++ cat /tmp/tmp.lzS4XVzmSR ++ rm /tmp/tmp.3wi7p5EG7v /tmp/tmp.lzS4XVzmSR ++ return 0 + secret_pass='7OVnIPdog&]S>X-4' ++ getSecretData internal-some-name root ++ local secretName=internal-some-name ++ local dataKey=root ++ kubectl_bin get secrets/internal-some-name '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pmxbix8gyR +++ mktemp ++ local LAST_ERR=/tmp/tmp.sgP0dRcqWq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Pmxbix8gyR ++ cat /tmp/tmp.sgP0dRcqWq ++ rm /tmp/tmp.Pmxbix8gyR /tmp/tmp.sgP0dRcqWq ++ return 0 + int_secret_pass='7OVnIPdog&]S>X-4' + [[ -z 7OVnIPdog&]S>X-4 ]] + [[ 7OVnIPdog&]S>X-4 != \7\O\V\n\I\P\d\o\g\&\]\S\>\X\-\4 ]] + [[ root != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ root ]] + [[ '' =~ root ]] + echo 'Running compare for root' Running compare for root + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''7OVnIPdog&]S>X-4'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S87Gb09sPF +++ mktemp ++ local LAST_ERR=/tmp/tmp.bXyai0KIrB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S87Gb09sPF ++ cat /tmp/tmp.bXyai0KIrB ++ rm /tmp/tmp.S87Gb09sPF /tmp/tmp.bXyai0KIrB ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking xtrabackup' Checking xtrabackup ++ getSecretData my-cluster-secrets xtrabackup ++ local secretName=my-cluster-secrets ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.nkrA0uQEm2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zOGxIXJKBB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nkrA0uQEm2 ++ cat /tmp/tmp.zOGxIXJKBB ++ rm /tmp/tmp.nkrA0uQEm2 /tmp/tmp.zOGxIXJKBB ++ return 0 + secret_pass='7ku@1UhR+2emw2Y%,<' ++ getSecretData internal-some-name xtrabackup ++ local secretName=internal-some-name ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.h9xbvmHfD1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SwMjlDr2gN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h9xbvmHfD1 ++ cat /tmp/tmp.SwMjlDr2gN ++ rm /tmp/tmp.h9xbvmHfD1 /tmp/tmp.SwMjlDr2gN ++ return 0 + int_secret_pass='7ku@1UhR+2emw2Y%,<' + [[ -z 7ku@1UhR+2emw2Y%,< ]] + [[ 7ku@1UhR+2emw2Y%,< != \7\k\u\@\1\U\h\R\+\2\e\m\w\2\Y\%\,\< ]] + [[ xtrabackup != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ xtrabackup ]] + [[ '' =~ xtrabackup ]] + echo 'Running compare for xtrabackup' Running compare for xtrabackup + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\''7ku@1UhR+2emw2Y%,<'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\''7ku@1UhR+2emw2Y%,<'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\''7ku@1UhR+2emw2Y%,<'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\''7ku@1UhR+2emw2Y%,<'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v0iJUtMCMH +++ mktemp ++ local LAST_ERR=/tmp/tmp.bYoOge66Jr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v0iJUtMCMH ++ cat /tmp/tmp.bYoOge66Jr ++ rm /tmp/tmp.v0iJUtMCMH /tmp/tmp.bYoOge66Jr ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking monitor' Checking monitor ++ getSecretData my-cluster-secrets monitor ++ local secretName=my-cluster-secrets ++ local dataKey=monitor ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.7ZlS2BGvou +++ mktemp ++ local LAST_ERR=/tmp/tmp.0K5pWRQFqs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7ZlS2BGvou ++ cat /tmp/tmp.0K5pWRQFqs ++ rm /tmp/tmp.7ZlS2BGvou /tmp/tmp.0K5pWRQFqs ++ return 0 + secret_pass=monitor_password ++ getSecretData internal-some-name monitor ++ local secretName=internal-some-name ++ local dataKey=monitor ++ kubectl_bin get secrets/internal-some-name '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.2gmOiCyBE4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DsXfJYCBUH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2gmOiCyBE4 ++ cat /tmp/tmp.DsXfJYCBUH ++ rm /tmp/tmp.2gmOiCyBE4 /tmp/tmp.DsXfJYCBUH ++ return 0 + int_secret_pass=monitor_password + [[ -z monitor_password ]] + [[ monitor_password != \m\o\n\i\t\o\r\_\p\a\s\s\w\o\r\d ]] + [[ monitor != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ monitor ]] + [[ '' =~ monitor ]] + echo 'Running compare for monitor' Running compare for monitor + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mq8QKUAMtJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.R6XX8qk3cU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mq8QKUAMtJ ++ cat /tmp/tmp.R6XX8qk3cU ++ rm /tmp/tmp.mq8QKUAMtJ /tmp/tmp.R6XX8qk3cU ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking proxyadmin' Checking proxyadmin ++ getSecretData my-cluster-secrets proxyadmin ++ local secretName=my-cluster-secrets ++ local dataKey=proxyadmin ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.Tf0GYM2Wrr +++ mktemp ++ local LAST_ERR=/tmp/tmp.GOpGHClmjl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Tf0GYM2Wrr ++ cat /tmp/tmp.GOpGHClmjl ++ rm /tmp/tmp.Tf0GYM2Wrr /tmp/tmp.GOpGHClmjl ++ return 0 + secret_pass='5)Pz]$gD@~O+^6$I>U' ++ getSecretData internal-some-name proxyadmin ++ local secretName=internal-some-name ++ local dataKey=proxyadmin ++ kubectl_bin get secrets/internal-some-name '--template={{.data.proxyadmin}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.6hIaHLu4uK +++ mktemp ++ local LAST_ERR=/tmp/tmp.FwXhChSk80 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.proxyadmin}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6hIaHLu4uK ++ cat /tmp/tmp.FwXhChSk80 ++ rm /tmp/tmp.6hIaHLu4uK /tmp/tmp.FwXhChSk80 ++ return 0 + int_secret_pass='5)Pz]$gD@~O+^6$I>U' + [[ -z 5)Pz]$gD@~O+^6$I>U ]] + [[ 5)Pz]$gD@~O+^6$I>U != \5\)\P\z\]\$\g\D\@\~\O\+\^\6\$\I\>\U ]] + [[ proxyadmin != \p\r\o\x\y\a\d\m\i\n ]] + [[ proxyadmin == \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ proxyadmin ]] + [[ '' =~ proxyadmin ]] + echo 'Running compare for proxyadmin' Running compare for proxyadmin + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''5)Pz]$gD@~O+^6$I>U'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''5)Pz]$gD@~O+^6$I>U'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''5)Pz]$gD@~O+^6$I>U'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''5)Pz]$gD@~O+^6$I>U'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql /tmp/tmp.miV4rrH5zD/select-2.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking operator' Checking operator ++ getSecretData my-cluster-secrets operator ++ local secretName=my-cluster-secrets ++ local dataKey=operator ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.4am6faywZg +++ mktemp ++ local LAST_ERR=/tmp/tmp.CmYkhbjhPw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4am6faywZg ++ cat /tmp/tmp.CmYkhbjhPw ++ rm /tmp/tmp.4am6faywZg /tmp/tmp.CmYkhbjhPw ++ return 0 + secret_pass='%6oP)DrIG_5}g*j!-' ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.bYGtmKKMaU +++ mktemp ++ local LAST_ERR=/tmp/tmp.huhpaOFbrl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bYGtmKKMaU ++ cat /tmp/tmp.huhpaOFbrl ++ rm /tmp/tmp.bYGtmKKMaU /tmp/tmp.huhpaOFbrl ++ return 0 + int_secret_pass='%6oP)DrIG_5}g*j!-' + [[ -z %6oP)DrIG_5}g*j!- ]] + [[ %6oP)DrIG_5}g*j!- != \%\6\o\P\)\D\r\I\G\_\5\}\g\*\j\!\- ]] + [[ operator != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ operator ]] + [[ '' =~ operator ]] + echo 'Running compare for operator' Running compare for operator + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''%6oP)DrIG_5}g*j!-'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''%6oP)DrIG_5}g*j!-'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''%6oP)DrIG_5}g*j!-'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''%6oP)DrIG_5}g*j!-'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cuGmiy5oau +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZUSFxH9pQ2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cuGmiy5oau ++ cat /tmp/tmp.ZUSFxH9pQ2 ++ rm /tmp/tmp.cuGmiy5oau /tmp/tmp.ZUSFxH9pQ2 ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking replication' Checking replication ++ getSecretData my-cluster-secrets replication ++ local secretName=my-cluster-secrets ++ local dataKey=replication ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.replication}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.qP7hX3zopT +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z8IjDZn6oa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.replication}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qP7hX3zopT ++ cat /tmp/tmp.Z8IjDZn6oa ++ rm /tmp/tmp.qP7hX3zopT /tmp/tmp.Z8IjDZn6oa ++ return 0 + secret_pass='SA,]icW{zzXEEGOTk9e' ++ getSecretData internal-some-name replication ++ local secretName=internal-some-name ++ local dataKey=replication ++ kubectl_bin get secrets/internal-some-name '--template={{.data.replication}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.Alz9nQKJfu +++ mktemp ++ local LAST_ERR=/tmp/tmp.COVqnKlAXh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.replication}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Alz9nQKJfu ++ cat /tmp/tmp.COVqnKlAXh ++ rm /tmp/tmp.Alz9nQKJfu /tmp/tmp.COVqnKlAXh ++ return 0 + int_secret_pass='SA,]icW{zzXEEGOTk9e' + [[ -z SA,]icW{zzXEEGOTk9e ]] + [[ SA,]icW{zzXEEGOTk9e != \S\A\,\]\i\c\W\{\z\z\X\E\E\G\O\T\k\9\e ]] + [[ replication != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ replication ]] + [[ '' =~ replication ]] + echo 'Running compare for replication' Running compare for replication + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\''SA,]icW{zzXEEGOTk9e'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\''SA,]icW{zzXEEGOTk9e'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\''SA,]icW{zzXEEGOTk9e'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\''SA,]icW{zzXEEGOTk9e'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vb5PHJsgM5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cSqPmS1FlQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vb5PHJsgM5 ++ cat /tmp/tmp.cSqPmS1FlQ ++ rm /tmp/tmp.vb5PHJsgM5 /tmp/tmp.cSqPmS1FlQ ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + [[ -n '' ]] + [[ -n '' ]] + desc 'test root' + set +o xtrace ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets root dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=root + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.YFpPVJUa32 ++ mktemp + local LAST_ERR=/tmp/tmp.gSW7FjxtXw + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YFpPVJUa32 secret/my-cluster-secrets patched + cat /tmp/tmp.gSW7FjxtXw + rm /tmp/tmp.YFpPVJUa32 /tmp/tmp.gSW7FjxtXw + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dEffADHlY3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.5jpIbtggLt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dEffADHlY3 ++ cat /tmp/tmp.5jpIbtggLt ++ rm /tmp/tmp.dEffADHlY3 /tmp/tmp.5jpIbtggLt ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + desc 'test proxyadmin' + set +o xtrace ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.cusXDZTyVu ++ mktemp + local LAST_ERR=/tmp/tmp.OBoOpWyn6i + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cusXDZTyVu perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.OBoOpWyn6i + rm /tmp/tmp.cusXDZTyVu /tmp/tmp.OBoOpWyn6i + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Jv7uKzDgg5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PEI9b8soYZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Jv7uKzDgg5 ++ cat /tmp/tmp.PEI9b8soYZ ++ rm /tmp/tmp.Jv7uKzDgg5 /tmp/tmp.PEI9b8soYZ ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C44mJtm0qX +++ mktemp ++ local LAST_ERR=/tmp/tmp.LRCxbt9Iky ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.C44mJtm0qX ++ cat /tmp/tmp.LRCxbt9Iky ++ rm /tmp/tmp.C44mJtm0qX /tmp/tmp.LRCxbt9Iky ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.L9qDJ6luMs ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.tykRzjTGUo +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.L9qDJ6luMs +++++ cat /tmp/tmp.tykRzjTGUo +++++ rm /tmp/tmp.L9qDJ6luMs /tmp/tmp.tykRzjTGUo +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Y7ChMvDRBG ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.cZYob53H1y +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Y7ChMvDRBG +++++ cat /tmp/tmp.cZYob53H1y +++++ rm /tmp/tmp.Y7ChMvDRBG /tmp/tmp.cZYob53H1y +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZfjnxzVga1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FNq1y93CEh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZfjnxzVga1 ++ cat /tmp/tmp.FNq1y93CEh ++ rm /tmp/tmp.ZfjnxzVga1 /tmp/tmp.FNq1y93CEh ++ return 0 + [[ 3 == \3 ]] + echo + patch_secret my-cluster-secrets proxyadmin dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=proxyadmin + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.wxkhcGFFCz ++ mktemp + local LAST_ERR=/tmp/tmp.kJUm5Kmf6z + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wxkhcGFFCz secret/my-cluster-secrets patched + cat /tmp/tmp.kJUm5Kmf6z + rm /tmp/tmp.wxkhcGFFCz /tmp/tmp.kJUm5Kmf6z + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nd1tP8oGob +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZXxo7VXpzz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nd1tP8oGob ++ cat /tmp/tmp.ZXxo7VXpzz ++ rm /tmp/tmp.nd1tP8oGob /tmp/tmp.ZXxo7VXpzz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dlnVVRoj78 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9SJCY21fKd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dlnVVRoj78 ++ cat /tmp/tmp.9SJCY21fKd ++ rm /tmp/tmp.dlnVVRoj78 /tmp/tmp.9SJCY21fKd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qaSNHA2zMX +++ mktemp ++ local LAST_ERR=/tmp/tmp.4Bentdnzgx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qaSNHA2zMX ++ cat /tmp/tmp.4Bentdnzgx ++ rm /tmp/tmp.qaSNHA2zMX /tmp/tmp.4Bentdnzgx ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6PHTbusUfw +++ mktemp ++ local LAST_ERR=/tmp/tmp.tFRa21AzIT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6PHTbusUfw ++ cat /tmp/tmp.tFRa21AzIT ++ rm /tmp/tmp.6PHTbusUfw /tmp/tmp.tFRa21AzIT ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.K6SLvVGRcK ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.rEtususkyR +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.K6SLvVGRcK +++++ cat /tmp/tmp.rEtususkyR +++++ rm /tmp/tmp.K6SLvVGRcK /tmp/tmp.rEtususkyR +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.T5Rd4IkVYp ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.tic0lfDxk0 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.T5Rd4IkVYp +++++ cat /tmp/tmp.tic0lfDxk0 +++++ rm /tmp/tmp.T5Rd4IkVYp /tmp/tmp.tic0lfDxk0 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0bAm2fLMlJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.xpxppQudnT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0bAm2fLMlJ ++ cat /tmp/tmp.xpxppQudnT ++ rm /tmp/tmp.0bAm2fLMlJ /tmp/tmp.xpxppQudnT ++ return 0 + [[ 3 == \3 ]] + echo + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql /tmp/tmp.miV4rrH5zD/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql /tmp/tmp.miV4rrH5zD/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-2.sql /tmp/tmp.miV4rrH5zD/select-2.sql + desc 'test xtrabackup' + set +o xtrace ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.FjJUxvqEXg ++ mktemp + local LAST_ERR=/tmp/tmp.W16nzfWAkJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FjJUxvqEXg perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.W16nzfWAkJ + rm /tmp/tmp.FjJUxvqEXg /tmp/tmp.W16nzfWAkJ + return 0 + patch_secret my-cluster-secrets xtrabackup dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=xtrabackup + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.BKz1zTHNR6 ++ mktemp + local LAST_ERR=/tmp/tmp.VeOawsBf9u + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BKz1zTHNR6 secret/my-cluster-secrets patched + cat /tmp/tmp.VeOawsBf9u + rm /tmp/tmp.BKz1zTHNR6 /tmp/tmp.VeOawsBf9u + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.16zZd7XQwo +++ mktemp ++ local LAST_ERR=/tmp/tmp.05tEM6hbPT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.16zZd7XQwo ++ cat /tmp/tmp.05tEM6hbPT ++ rm /tmp/tmp.16zZd7XQwo /tmp/tmp.05tEM6hbPT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JlBpPLPIUY +++ mktemp ++ local LAST_ERR=/tmp/tmp.FOjv58xRQy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JlBpPLPIUY ++ cat /tmp/tmp.FOjv58xRQy ++ rm /tmp/tmp.JlBpPLPIUY /tmp/tmp.FOjv58xRQy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XE7r6W2al2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.n0FIuxu5CF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XE7r6W2al2 ++ cat /tmp/tmp.n0FIuxu5CF ++ rm /tmp/tmp.XE7r6W2al2 /tmp/tmp.n0FIuxu5CF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.soPMv57uZV +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZBvbaytb65 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.soPMv57uZV ++ cat /tmp/tmp.ZBvbaytb65 ++ rm /tmp/tmp.soPMv57uZV /tmp/tmp.ZBvbaytb65 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Q02Wwzf5lR +++ mktemp ++ local LAST_ERR=/tmp/tmp.qMGqbvGOHd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Q02Wwzf5lR ++ cat /tmp/tmp.qMGqbvGOHd ++ rm /tmp/tmp.Q02Wwzf5lR /tmp/tmp.qMGqbvGOHd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k3pUfAc7in +++ mktemp ++ local LAST_ERR=/tmp/tmp.mWKnLphH9K ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k3pUfAc7in ++ cat /tmp/tmp.mWKnLphH9K ++ rm /tmp/tmp.k3pUfAc7in /tmp/tmp.mWKnLphH9K ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9W3O5U9Svk +++ mktemp ++ local LAST_ERR=/tmp/tmp.3Y7Upk2Gge ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9W3O5U9Svk ++ cat /tmp/tmp.3Y7Upk2Gge ++ rm /tmp/tmp.9W3O5U9Svk /tmp/tmp.3Y7Upk2Gge ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lP7ebFIkMO +++ mktemp ++ local LAST_ERR=/tmp/tmp.4kG7PbF2cj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lP7ebFIkMO ++ cat /tmp/tmp.4kG7PbF2cj ++ rm /tmp/tmp.lP7ebFIkMO /tmp/tmp.4kG7PbF2cj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4SEoYOPrYq +++ mktemp ++ local LAST_ERR=/tmp/tmp.SQMXCaQimm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4SEoYOPrYq ++ cat /tmp/tmp.SQMXCaQimm ++ rm /tmp/tmp.4SEoYOPrYq /tmp/tmp.SQMXCaQimm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dc7Bz8W9fK +++ mktemp ++ local LAST_ERR=/tmp/tmp.2LA3ifbzsD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dc7Bz8W9fK ++ cat /tmp/tmp.2LA3ifbzsD ++ rm /tmp/tmp.dc7Bz8W9fK /tmp/tmp.2LA3ifbzsD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xxdb7ZEIGA +++ mktemp ++ local LAST_ERR=/tmp/tmp.6GuvOjw3GT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xxdb7ZEIGA ++ cat /tmp/tmp.6GuvOjw3GT ++ rm /tmp/tmp.Xxdb7ZEIGA /tmp/tmp.6GuvOjw3GT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LhgTYnWhkn +++ mktemp ++ local LAST_ERR=/tmp/tmp.weL0nfuBQG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LhgTYnWhkn ++ cat /tmp/tmp.weL0nfuBQG ++ rm /tmp/tmp.LhgTYnWhkn /tmp/tmp.weL0nfuBQG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kxx24qQIja +++ mktemp ++ local LAST_ERR=/tmp/tmp.zXOi3vatAQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kxx24qQIja ++ cat /tmp/tmp.zXOi3vatAQ ++ rm /tmp/tmp.kxx24qQIja /tmp/tmp.zXOi3vatAQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Nos65PIvAm +++ mktemp ++ local LAST_ERR=/tmp/tmp.rHQU3TjFDp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Nos65PIvAm ++ cat /tmp/tmp.rHQU3TjFDp ++ rm /tmp/tmp.Nos65PIvAm /tmp/tmp.rHQU3TjFDp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gjEI0zu7GJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.60wXeyOU67 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gjEI0zu7GJ ++ cat /tmp/tmp.60wXeyOU67 ++ rm /tmp/tmp.gjEI0zu7GJ /tmp/tmp.60wXeyOU67 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BqXTFgrwh0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tUZtPGDRMF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BqXTFgrwh0 ++ cat /tmp/tmp.tUZtPGDRMF ++ rm /tmp/tmp.BqXTFgrwh0 /tmp/tmp.tUZtPGDRMF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QrQ93kvxp6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.NkUS6JW75F ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QrQ93kvxp6 ++ cat /tmp/tmp.NkUS6JW75F ++ rm /tmp/tmp.QrQ93kvxp6 /tmp/tmp.NkUS6JW75F ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ySAIKd9o6x +++ mktemp ++ local LAST_ERR=/tmp/tmp.vMRmm3M3Bu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ySAIKd9o6x ++ cat /tmp/tmp.vMRmm3M3Bu ++ rm /tmp/tmp.ySAIKd9o6x /tmp/tmp.vMRmm3M3Bu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PYeFsJEGCG +++ mktemp ++ local LAST_ERR=/tmp/tmp.0Imqcf05sj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PYeFsJEGCG ++ cat /tmp/tmp.0Imqcf05sj ++ rm /tmp/tmp.PYeFsJEGCG /tmp/tmp.0Imqcf05sj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mSRuDucT62 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qy2687FMVC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mSRuDucT62 ++ cat /tmp/tmp.qy2687FMVC ++ rm /tmp/tmp.mSRuDucT62 /tmp/tmp.qy2687FMVC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rx76aTvlZi +++ mktemp ++ local LAST_ERR=/tmp/tmp.dUblIeaNRm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rx76aTvlZi ++ cat /tmp/tmp.dUblIeaNRm ++ rm /tmp/tmp.rx76aTvlZi /tmp/tmp.dUblIeaNRm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MhzUtMWenH +++ mktemp ++ local LAST_ERR=/tmp/tmp.N96Vq0yqf3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MhzUtMWenH ++ cat /tmp/tmp.N96Vq0yqf3 ++ rm /tmp/tmp.MhzUtMWenH /tmp/tmp.N96Vq0yqf3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HO3I57n6fc +++ mktemp ++ local LAST_ERR=/tmp/tmp.e1KZj6KA0H ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HO3I57n6fc ++ cat /tmp/tmp.e1KZj6KA0H ++ rm /tmp/tmp.HO3I57n6fc /tmp/tmp.e1KZj6KA0H ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d3u79i4Ug5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.nrRgHjIIuI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.d3u79i4Ug5 ++ cat /tmp/tmp.nrRgHjIIuI ++ rm /tmp/tmp.d3u79i4Ug5 /tmp/tmp.nrRgHjIIuI ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cCOqIs6gsK +++ mktemp ++ local LAST_ERR=/tmp/tmp.HB2J4y1SOu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cCOqIs6gsK ++ cat /tmp/tmp.HB2J4y1SOu ++ rm /tmp/tmp.cCOqIs6gsK /tmp/tmp.HB2J4y1SOu ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.iE8FMIjLhR ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.kc4xBfg7fz +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.iE8FMIjLhR +++++ cat /tmp/tmp.kc4xBfg7fz +++++ rm /tmp/tmp.iE8FMIjLhR /tmp/tmp.kc4xBfg7fz +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.5oRhGu4jJA ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.p1C0PWHg0i +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.5oRhGu4jJA +++++ cat /tmp/tmp.p1C0PWHg0i +++++ rm /tmp/tmp.5oRhGu4jJA /tmp/tmp.p1C0PWHg0i +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4DcZ1UyQPo +++ mktemp ++ local LAST_ERR=/tmp/tmp.xwQ74zImpj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4DcZ1UyQPo ++ cat /tmp/tmp.xwQ74zImpj ++ rm /tmp/tmp.4DcZ1UyQPo /tmp/tmp.xwQ74zImpj ++ return 0 + [[ 2 == \2 ]] + echo + compare_mysql_cmd_local select-3 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 '' pxc + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 pxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-3.sql /tmp/tmp.miV4rrH5zD/select-3.sql + desc 'test monitor' + set +o xtrace ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.CKYs3BFcxg ++ mktemp + local LAST_ERR=/tmp/tmp.fBcXBJqEQX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CKYs3BFcxg secret/my-cluster-secrets patched + cat /tmp/tmp.fBcXBJqEQX + rm /tmp/tmp.CKYs3BFcxg /tmp/tmp.fBcXBJqEQX + return 0 + wait_for_password_propagation my-cluster-secrets monitor + local secret=my-cluster-secrets + local user=monitor + local max_retry=240 ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.pfmpjIrCOF +++ mktemp ++ local LAST_ERR=/tmp/tmp.OzKLhBnhiG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pfmpjIrCOF ++ cat /tmp/tmp.OzKLhBnhiG ++ rm /tmp/tmp.pfmpjIrCOF /tmp/tmp.OzKLhBnhiG ++ return 0 + local root_pass=test-password + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + echo 'Skipping dual password feature doesn'\''t work for 5.7. PXC 5.7 doesn'\''t support it!' Skipping dual password feature doesn't work for 5.7. PXC 5.7 doesn't support it! + return + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EBMa79r4Ux +++ mktemp ++ local LAST_ERR=/tmp/tmp.g30VuZ3aJE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EBMa79r4Ux ++ cat /tmp/tmp.g30VuZ3aJE ++ rm /tmp/tmp.EBMa79r4Ux /tmp/tmp.g30VuZ3aJE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PEr9KjSG02 +++ mktemp ++ local LAST_ERR=/tmp/tmp.mGq5ZlMkB0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PEr9KjSG02 ++ cat /tmp/tmp.mGq5ZlMkB0 ++ rm /tmp/tmp.PEr9KjSG02 /tmp/tmp.mGq5ZlMkB0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gUO4uxLdZM +++ mktemp ++ local LAST_ERR=/tmp/tmp.mXgjQNBawu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gUO4uxLdZM ++ cat /tmp/tmp.mXgjQNBawu ++ rm /tmp/tmp.gUO4uxLdZM /tmp/tmp.mXgjQNBawu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nwXhzsfYIm +++ mktemp ++ local LAST_ERR=/tmp/tmp.NXOEquExPP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nwXhzsfYIm ++ cat /tmp/tmp.NXOEquExPP ++ rm /tmp/tmp.nwXhzsfYIm /tmp/tmp.NXOEquExPP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XlVoscauJL +++ mktemp ++ local LAST_ERR=/tmp/tmp.eGer5VkQEA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XlVoscauJL ++ cat /tmp/tmp.eGer5VkQEA ++ rm /tmp/tmp.XlVoscauJL /tmp/tmp.eGer5VkQEA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.saYDbfOEyd +++ mktemp ++ local LAST_ERR=/tmp/tmp.4VzwJ12KXT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.saYDbfOEyd ++ cat /tmp/tmp.4VzwJ12KXT ++ rm /tmp/tmp.saYDbfOEyd /tmp/tmp.4VzwJ12KXT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dwd0iJ5OsN +++ mktemp ++ local LAST_ERR=/tmp/tmp.aEPP4W6yPK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Dwd0iJ5OsN ++ cat /tmp/tmp.aEPP4W6yPK ++ rm /tmp/tmp.Dwd0iJ5OsN /tmp/tmp.aEPP4W6yPK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jjKFZXVMWz +++ mktemp ++ local LAST_ERR=/tmp/tmp.36ad8Gyyzt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jjKFZXVMWz ++ cat /tmp/tmp.36ad8Gyyzt ++ rm /tmp/tmp.jjKFZXVMWz /tmp/tmp.36ad8Gyyzt ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JO0GOmRrDf +++ mktemp ++ local LAST_ERR=/tmp/tmp.zA6OiL15rn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JO0GOmRrDf ++ cat /tmp/tmp.zA6OiL15rn ++ rm /tmp/tmp.JO0GOmRrDf /tmp/tmp.zA6OiL15rn ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.geP88Aib1h ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.uCbC7jdWED +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.geP88Aib1h +++++ cat /tmp/tmp.uCbC7jdWED +++++ rm /tmp/tmp.geP88Aib1h /tmp/tmp.uCbC7jdWED +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.gRSWEGoXAa ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.r6EuAl351B +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.gRSWEGoXAa +++++ cat /tmp/tmp.r6EuAl351B +++++ rm /tmp/tmp.gRSWEGoXAa /tmp/tmp.r6EuAl351B +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.b2pNhQX0I1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qma5kZsFle ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.b2pNhQX0I1 ++ cat /tmp/tmp.Qma5kZsFle ++ rm /tmp/tmp.b2pNhQX0I1 /tmp/tmp.Qma5kZsFle ++ return 0 + [[ 2 == \2 ]] + echo + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iWTC36JyQ1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.fxtieOuLLV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iWTC36JyQ1 ++ cat /tmp/tmp.fxtieOuLLV ++ rm /tmp/tmp.iWTC36JyQ1 /tmp/tmp.fxtieOuLLV ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + desc 'test operator' + set +o xtrace ----------------------------------------------------------------------------------- test operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets operator dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=operator + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.Etl6FT72Tq ++ mktemp + local LAST_ERR=/tmp/tmp.AXDhzwC9ix + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Etl6FT72Tq secret/my-cluster-secrets patched + cat /tmp/tmp.AXDhzwC9ix + rm /tmp/tmp.Etl6FT72Tq /tmp/tmp.AXDhzwC9ix + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JN0JzEwq9p +++ mktemp ++ local LAST_ERR=/tmp/tmp.CnPOItZ6hz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JN0JzEwq9p ++ cat /tmp/tmp.CnPOItZ6hz ++ rm /tmp/tmp.JN0JzEwq9p /tmp/tmp.CnPOItZ6hz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rErQpQDJ92 +++ mktemp ++ local LAST_ERR=/tmp/tmp.p0eOwuGTWX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rErQpQDJ92 ++ cat /tmp/tmp.p0eOwuGTWX ++ rm /tmp/tmp.rErQpQDJ92 /tmp/tmp.p0eOwuGTWX ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.X76qUaExEz +++ mktemp ++ local LAST_ERR=/tmp/tmp.tW7plLPjt0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.X76qUaExEz ++ cat /tmp/tmp.tW7plLPjt0 ++ rm /tmp/tmp.X76qUaExEz /tmp/tmp.tW7plLPjt0 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.nSzFXizt3q ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.p4WrIaT6rb +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.nSzFXizt3q +++++ cat /tmp/tmp.p4WrIaT6rb +++++ rm /tmp/tmp.nSzFXizt3q /tmp/tmp.p4WrIaT6rb +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.9mBO2xCtyv ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.OEjLhNhFea +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.9mBO2xCtyv +++++ cat /tmp/tmp.OEjLhNhFea +++++ rm /tmp/tmp.9mBO2xCtyv /tmp/tmp.OEjLhNhFea +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GtcJgCK6N4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.r16y2SFWdT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GtcJgCK6N4 ++ cat /tmp/tmp.r16y2SFWdT ++ rm /tmp/tmp.GtcJgCK6N4 /tmp/tmp.r16y2SFWdT ++ return 0 + [[ 2 == \2 ]] + echo + sleep 10 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dnGYzqOQBd +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z270f5Lwsj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dnGYzqOQBd ++ cat /tmp/tmp.Z270f5Lwsj ++ rm /tmp/tmp.dnGYzqOQBd /tmp/tmp.Z270f5Lwsj ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + desc 'change secret name' + set +o xtrace ----------------------------------------------------------------------------------- change secret name ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.wkyzXFAU7Q ++ mktemp + local LAST_ERR=/tmp/tmp.cJqZNTfa4F + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wkyzXFAU7Q perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.cJqZNTfa4F + rm /tmp/tmp.wkyzXFAU7Q /tmp/tmp.cJqZNTfa4F + return 0 + sleep 30 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.12Dh28detz +++ mktemp ++ local LAST_ERR=/tmp/tmp.tqKiT6zzLL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.12Dh28detz ++ cat /tmp/tmp.tqKiT6zzLL ++ rm /tmp/tmp.12Dh28detz /tmp/tmp.tqKiT6zzLL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UvrJ6EaQn5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.uMgI7fxa0c ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UvrJ6EaQn5 ++ cat /tmp/tmp.uMgI7fxa0c ++ rm /tmp/tmp.UvrJ6EaQn5 /tmp/tmp.uMgI7fxa0c ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w54UA5wuhi +++ mktemp ++ local LAST_ERR=/tmp/tmp.LuD9dRMlPB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.w54UA5wuhi ++ cat /tmp/tmp.LuD9dRMlPB ++ rm /tmp/tmp.w54UA5wuhi /tmp/tmp.LuD9dRMlPB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sOjrSzXyE3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.NKe4K2wOcC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sOjrSzXyE3 ++ cat /tmp/tmp.NKe4K2wOcC ++ rm /tmp/tmp.sOjrSzXyE3 /tmp/tmp.NKe4K2wOcC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UFTtHDgcuh +++ mktemp ++ local LAST_ERR=/tmp/tmp.cjj5g88ZPm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UFTtHDgcuh ++ cat /tmp/tmp.cjj5g88ZPm ++ rm /tmp/tmp.UFTtHDgcuh /tmp/tmp.cjj5g88ZPm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aC3szhlLZW +++ mktemp ++ local LAST_ERR=/tmp/tmp.XY5EgkHnf7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aC3szhlLZW ++ cat /tmp/tmp.XY5EgkHnf7 ++ rm /tmp/tmp.aC3szhlLZW /tmp/tmp.XY5EgkHnf7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2soE0AZWfL +++ mktemp ++ local LAST_ERR=/tmp/tmp.7wiDmWGoVP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2soE0AZWfL ++ cat /tmp/tmp.7wiDmWGoVP ++ rm /tmp/tmp.2soE0AZWfL /tmp/tmp.7wiDmWGoVP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RHH6E8uby8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.z1meAHWuUp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RHH6E8uby8 ++ cat /tmp/tmp.z1meAHWuUp ++ rm /tmp/tmp.RHH6E8uby8 /tmp/tmp.z1meAHWuUp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iTivWHDGzy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ph9cQldaI7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iTivWHDGzy ++ cat /tmp/tmp.Ph9cQldaI7 ++ rm /tmp/tmp.iTivWHDGzy /tmp/tmp.Ph9cQldaI7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IVEP2khbB0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Xrr4oZcoU3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IVEP2khbB0 ++ cat /tmp/tmp.Xrr4oZcoU3 ++ rm /tmp/tmp.IVEP2khbB0 /tmp/tmp.Xrr4oZcoU3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3FoOCY1qWL +++ mktemp ++ local LAST_ERR=/tmp/tmp.804jR3oc1E ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3FoOCY1qWL ++ cat /tmp/tmp.804jR3oc1E ++ rm /tmp/tmp.3FoOCY1qWL /tmp/tmp.804jR3oc1E ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Atux2NaBJI +++ mktemp ++ local LAST_ERR=/tmp/tmp.DRgVyIH4M7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Atux2NaBJI ++ cat /tmp/tmp.DRgVyIH4M7 ++ rm /tmp/tmp.Atux2NaBJI /tmp/tmp.DRgVyIH4M7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z3qgxny50W +++ mktemp ++ local LAST_ERR=/tmp/tmp.dXfm6UebEW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Z3qgxny50W ++ cat /tmp/tmp.dXfm6UebEW ++ rm /tmp/tmp.Z3qgxny50W /tmp/tmp.dXfm6UebEW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.n9LvRHf7it +++ mktemp ++ local LAST_ERR=/tmp/tmp.nf3SG5weiA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.n9LvRHf7it ++ cat /tmp/tmp.nf3SG5weiA ++ rm /tmp/tmp.n9LvRHf7it /tmp/tmp.nf3SG5weiA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NCYolBaYFd +++ mktemp ++ local LAST_ERR=/tmp/tmp.076U3JVDTQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NCYolBaYFd ++ cat /tmp/tmp.076U3JVDTQ ++ rm /tmp/tmp.NCYolBaYFd /tmp/tmp.076U3JVDTQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.liGQZykMPU +++ mktemp ++ local LAST_ERR=/tmp/tmp.HKRK4gOvW8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.liGQZykMPU ++ cat /tmp/tmp.HKRK4gOvW8 ++ rm /tmp/tmp.liGQZykMPU /tmp/tmp.HKRK4gOvW8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aIXaYvSkw3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.j6hjjrLXtF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aIXaYvSkw3 ++ cat /tmp/tmp.j6hjjrLXtF ++ rm /tmp/tmp.aIXaYvSkw3 /tmp/tmp.j6hjjrLXtF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6LyzGcFuDg +++ mktemp ++ local LAST_ERR=/tmp/tmp.hKTf8ixBMW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6LyzGcFuDg ++ cat /tmp/tmp.hKTf8ixBMW ++ rm /tmp/tmp.6LyzGcFuDg /tmp/tmp.hKTf8ixBMW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.55WhXf0Ajr +++ mktemp ++ local LAST_ERR=/tmp/tmp.XuAzB3d1hS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.55WhXf0Ajr ++ cat /tmp/tmp.XuAzB3d1hS ++ rm /tmp/tmp.55WhXf0Ajr /tmp/tmp.XuAzB3d1hS ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qrTzRfDDfa +++ mktemp ++ local LAST_ERR=/tmp/tmp.kndB8FIc3A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qrTzRfDDfa ++ cat /tmp/tmp.kndB8FIc3A ++ rm /tmp/tmp.qrTzRfDDfa /tmp/tmp.kndB8FIc3A ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.oR3BojheRj ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.BQBnVxnpyu +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.oR3BojheRj +++++ cat /tmp/tmp.BQBnVxnpyu +++++ rm /tmp/tmp.oR3BojheRj /tmp/tmp.BQBnVxnpyu +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.DCkOEcbUVW ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.JnjyiWYKri +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.DCkOEcbUVW +++++ cat /tmp/tmp.JnjyiWYKri +++++ rm /tmp/tmp.DCkOEcbUVW /tmp/tmp.JnjyiWYKri +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qIobbZDOY7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.sptVRcUGXv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qIobbZDOY7 ++ cat /tmp/tmp.sptVRcUGXv ++ rm /tmp/tmp.qIobbZDOY7 /tmp/tmp.sptVRcUGXv ++ return 0 + [[ 2 == \2 ]] + echo + desc 'test new operator' + set +o xtrace ----------------------------------------------------------------------------------- test new operator ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + patch_secret my-cluster-secrets-2 operator dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets-2 + local key=operator + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.WbK3bhZf5m ++ mktemp + local LAST_ERR=/tmp/tmp.AYhooWYuvv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WbK3bhZf5m secret/my-cluster-secrets-2 patched + cat /tmp/tmp.AYhooWYuvv + rm /tmp/tmp.WbK3bhZf5m /tmp/tmp.AYhooWYuvv + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Aw0LhMDOXg +++ mktemp ++ local LAST_ERR=/tmp/tmp.TQShCyVFiA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Aw0LhMDOXg ++ cat /tmp/tmp.TQShCyVFiA ++ rm /tmp/tmp.Aw0LhMDOXg /tmp/tmp.TQShCyVFiA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XO0QQbszhs +++ mktemp ++ local LAST_ERR=/tmp/tmp.vJ22aRZiCn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XO0QQbszhs ++ cat /tmp/tmp.vJ22aRZiCn ++ rm /tmp/tmp.XO0QQbszhs /tmp/tmp.vJ22aRZiCn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7ewCgOeHnY +++ mktemp ++ local LAST_ERR=/tmp/tmp.r3J4S99hCN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7ewCgOeHnY ++ cat /tmp/tmp.r3J4S99hCN ++ rm /tmp/tmp.7ewCgOeHnY /tmp/tmp.r3J4S99hCN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pd8oijJwBG +++ mktemp ++ local LAST_ERR=/tmp/tmp.Cd0dBXIYWE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pd8oijJwBG ++ cat /tmp/tmp.Cd0dBXIYWE ++ rm /tmp/tmp.pd8oijJwBG /tmp/tmp.Cd0dBXIYWE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.71yXQvI9EC +++ mktemp ++ local LAST_ERR=/tmp/tmp.uv0krA4dkR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.71yXQvI9EC ++ cat /tmp/tmp.uv0krA4dkR ++ rm /tmp/tmp.71yXQvI9EC /tmp/tmp.uv0krA4dkR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.n7joU21rz5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9nijEqOeC0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.n7joU21rz5 ++ cat /tmp/tmp.9nijEqOeC0 ++ rm /tmp/tmp.n7joU21rz5 /tmp/tmp.9nijEqOeC0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.smEaaOryw2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.o1Rc6InPvI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.smEaaOryw2 ++ cat /tmp/tmp.o1Rc6InPvI ++ rm /tmp/tmp.smEaaOryw2 /tmp/tmp.o1Rc6InPvI ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FCHa1FGOZl +++ mktemp ++ local LAST_ERR=/tmp/tmp.IXuhRnexz7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FCHa1FGOZl ++ cat /tmp/tmp.IXuhRnexz7 ++ rm /tmp/tmp.FCHa1FGOZl /tmp/tmp.IXuhRnexz7 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.7PBF0Ullh2 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.HzrEUSk5aI +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.7PBF0Ullh2 +++++ cat /tmp/tmp.HzrEUSk5aI +++++ rm /tmp/tmp.7PBF0Ullh2 /tmp/tmp.HzrEUSk5aI +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.sjCBQz3j7w ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.RYFFbh2uvQ +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.sjCBQz3j7w +++++ cat /tmp/tmp.RYFFbh2uvQ +++++ rm /tmp/tmp.sjCBQz3j7w /tmp/tmp.RYFFbh2uvQ +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.70bAm9C1si +++ mktemp ++ local LAST_ERR=/tmp/tmp.AFQMKsRP5Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.70bAm9C1si ++ cat /tmp/tmp.AFQMKsRP5Y ++ rm /tmp/tmp.70bAm9C1si /tmp/tmp.AFQMKsRP5Y ++ return 0 + [[ 2 == \2 ]] + echo + sleep 20 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v9cLblfNri +++ mktemp ++ local LAST_ERR=/tmp/tmp.M6cdCFGcct ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v9cLblfNri ++ cat /tmp/tmp.M6cdCFGcct ++ rm /tmp/tmp.v9cLblfNri /tmp/tmp.M6cdCFGcct ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql ++ getSecretData my-cluster-secrets-2 root ++ local secretName=my-cluster-secrets-2 ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.2BqUXCGip6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.NBCjpoBZKi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2BqUXCGip6 ++ cat /tmp/tmp.NBCjpoBZKi ++ rm /tmp/tmp.2BqUXCGip6 /tmp/tmp.NBCjpoBZKi ++ return 0 + newpass='nd^{k+lgS!D_C-M,ie' + desc 'test new users sync' + set +o xtrace ----------------------------------------------------------------------------------- test new users sync ----------------------------------------------------------------------------------- + run_mysql 'CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''nd^{k+lgS!D_C-M,ie'\'';' '-h some-name-pxc -uroot -p'\''nd^{k+lgS!D_C-M,ie'\''' + local 'command=CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''nd^{k+lgS!D_C-M,ie'\'';' + local 'uri=-h some-name-pxc -uroot -p'\''nd^{k+lgS!D_C-M,ie'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ajju1mY0lc +++ mktemp ++ local LAST_ERR=/tmp/tmp.WBoGiO1NQW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ajju1mY0lc ++ cat /tmp/tmp.WBoGiO1NQW ++ rm /tmp/tmp.Ajju1mY0lc /tmp/tmp.WBoGiO1NQW ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 40 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''nd^{k+lgS!D_C-M,ie'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''nd^{k+lgS!D_C-M,ie'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''nd^{k+lgS!D_C-M,ie'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''nd^{k+lgS!D_C-M,ie'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bOnqfenFSv +++ mktemp ++ local LAST_ERR=/tmp/tmp.hFEr6DriZf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bOnqfenFSv ++ cat /tmp/tmp.hFEr6DriZf ++ rm /tmp/tmp.bOnqfenFSv /tmp/tmp.hFEr6DriZf ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.l8C8iGO6BB +++ mktemp ++ local LAST_ERR=/tmp/tmp.RXgEDTZEHt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.l8C8iGO6BB ++ cat /tmp/tmp.RXgEDTZEHt ++ rm /tmp/tmp.l8C8iGO6BB /tmp/tmp.RXgEDTZEHt ++ return 0 + pass=test-password2 + desc 'check secret without operator' + set +o xtrace ----------------------------------------------------------------------------------- check secret without operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.cp65aOLpXl ++ mktemp + local LAST_ERR=/tmp/tmp.lgngI1fJ5V + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cp65aOLpXl secret/my-cluster-secrets-2 configured + cat /tmp/tmp.lgngI1fJ5V Warning: resource secrets/my-cluster-secrets-2 is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.cp65aOLpXl /tmp/tmp.lgngI1fJ5V + return 0 + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4-57.sql ]] + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WsMeEMO05z +++ mktemp ++ local LAST_ERR=/tmp/tmp.TcGtK6a1ku ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WsMeEMO05z ++ cat /tmp/tmp.TcGtK6a1ku ++ rm /tmp/tmp.WsMeEMO05z /tmp/tmp.TcGtK6a1ku ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-4.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-4.sql /tmp/tmp.miV4rrH5zD/select-4.sql + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/conf/some-name.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/conf/some-name.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/conf/some-name.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/conf/some-name.yml + local pvc_name= + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/conf/some-name.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + local LAST_OUT=/tmp/tmp.e6f388JB5M + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' ++ mktemp + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2232-31b0cef6#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + local LAST_ERR=/tmp/tmp.lQ0gQDIq1G + local exit_status=0 + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-3672~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' ++ seq 0 2 + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.e6f388JB5M perconaxtradbcluster.pxc.percona.com/some-name configured + cat /tmp/tmp.lQ0gQDIq1G + rm /tmp/tmp.e6f388JB5M /tmp/tmp.lQ0gQDIq1G + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i4i92mFER6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.x3o7705kfv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.i4i92mFER6 ++ cat /tmp/tmp.x3o7705kfv ++ rm /tmp/tmp.i4i92mFER6 /tmp/tmp.x3o7705kfv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FM3XpRvn1s +++ mktemp ++ local LAST_ERR=/tmp/tmp.IjmoRsEEyA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FM3XpRvn1s ++ cat /tmp/tmp.IjmoRsEEyA ++ rm /tmp/tmp.FM3XpRvn1s /tmp/tmp.IjmoRsEEyA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p4VaAeomZb +++ mktemp ++ local LAST_ERR=/tmp/tmp.522IG7VGYl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.p4VaAeomZb ++ cat /tmp/tmp.522IG7VGYl ++ rm /tmp/tmp.p4VaAeomZb /tmp/tmp.522IG7VGYl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KxBMtLpkgo +++ mktemp ++ local LAST_ERR=/tmp/tmp.XV7IuGhPPl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KxBMtLpkgo ++ cat /tmp/tmp.XV7IuGhPPl ++ rm /tmp/tmp.KxBMtLpkgo /tmp/tmp.XV7IuGhPPl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kDQbj3RAhh +++ mktemp ++ local LAST_ERR=/tmp/tmp.44INy6uQRZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kDQbj3RAhh ++ cat /tmp/tmp.44INy6uQRZ ++ rm /tmp/tmp.kDQbj3RAhh /tmp/tmp.44INy6uQRZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.piF6kgMiI7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.UKNqNhArm0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.piF6kgMiI7 ++ cat /tmp/tmp.UKNqNhArm0 ++ rm /tmp/tmp.piF6kgMiI7 /tmp/tmp.UKNqNhArm0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SUMZmInUVe +++ mktemp ++ local LAST_ERR=/tmp/tmp.Of8lu05so8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SUMZmInUVe ++ cat /tmp/tmp.Of8lu05so8 ++ rm /tmp/tmp.SUMZmInUVe /tmp/tmp.Of8lu05so8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fY5OBAxoBA +++ mktemp ++ local LAST_ERR=/tmp/tmp.x7G8uaW3eq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fY5OBAxoBA ++ cat /tmp/tmp.x7G8uaW3eq ++ rm /tmp/tmp.fY5OBAxoBA /tmp/tmp.x7G8uaW3eq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bCTIQaZetq +++ mktemp ++ local LAST_ERR=/tmp/tmp.AJgv54muNn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bCTIQaZetq ++ cat /tmp/tmp.AJgv54muNn ++ rm /tmp/tmp.bCTIQaZetq /tmp/tmp.AJgv54muNn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pu0ns5vFpP +++ mktemp ++ local LAST_ERR=/tmp/tmp.PJW4gO2qK7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Pu0ns5vFpP ++ cat /tmp/tmp.PJW4gO2qK7 ++ rm /tmp/tmp.Pu0ns5vFpP /tmp/tmp.PJW4gO2qK7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kxZo2Hx8lU +++ mktemp ++ local LAST_ERR=/tmp/tmp.Hw8F2ESh6l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kxZo2Hx8lU ++ cat /tmp/tmp.Hw8F2ESh6l ++ rm /tmp/tmp.kxZo2Hx8lU /tmp/tmp.Hw8F2ESh6l ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.losDvwSvBh +++ mktemp ++ local LAST_ERR=/tmp/tmp.dMMDhqYJeH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.losDvwSvBh ++ cat /tmp/tmp.dMMDhqYJeH ++ rm /tmp/tmp.losDvwSvBh /tmp/tmp.dMMDhqYJeH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MeoniF6Gfb +++ mktemp ++ local LAST_ERR=/tmp/tmp.JiSnur45Ex ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MeoniF6Gfb ++ cat /tmp/tmp.JiSnur45Ex ++ rm /tmp/tmp.MeoniF6Gfb /tmp/tmp.JiSnur45Ex ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V4TrMarDvL +++ mktemp ++ local LAST_ERR=/tmp/tmp.QeG20VMxgO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.V4TrMarDvL ++ cat /tmp/tmp.QeG20VMxgO ++ rm /tmp/tmp.V4TrMarDvL /tmp/tmp.QeG20VMxgO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MqAKaBAbYi +++ mktemp ++ local LAST_ERR=/tmp/tmp.jiRSzB3t1i ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MqAKaBAbYi ++ cat /tmp/tmp.jiRSzB3t1i ++ rm /tmp/tmp.MqAKaBAbYi /tmp/tmp.jiRSzB3t1i ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dqWBvDT5B4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PwsKODknZC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dqWBvDT5B4 ++ cat /tmp/tmp.PwsKODknZC ++ rm /tmp/tmp.dqWBvDT5B4 /tmp/tmp.PwsKODknZC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.J2kmMD2xyj +++ mktemp ++ local LAST_ERR=/tmp/tmp.hMTh7Ixh4l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.J2kmMD2xyj ++ cat /tmp/tmp.hMTh7Ixh4l ++ rm /tmp/tmp.J2kmMD2xyj /tmp/tmp.hMTh7Ixh4l ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Axj85K758S +++ mktemp ++ local LAST_ERR=/tmp/tmp.O0BxjYAEoN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Axj85K758S ++ cat /tmp/tmp.O0BxjYAEoN ++ rm /tmp/tmp.Axj85K758S /tmp/tmp.O0BxjYAEoN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YYhQfCxjVw +++ mktemp ++ local LAST_ERR=/tmp/tmp.JliQo4h58v ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YYhQfCxjVw ++ cat /tmp/tmp.JliQo4h58v ++ rm /tmp/tmp.YYhQfCxjVw /tmp/tmp.JliQo4h58v ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rm5tLkzM3I +++ mktemp ++ local LAST_ERR=/tmp/tmp.BED2ZN9hM5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rm5tLkzM3I ++ cat /tmp/tmp.BED2ZN9hM5 ++ rm /tmp/tmp.rm5tLkzM3I /tmp/tmp.BED2ZN9hM5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VCuDErmfNv +++ mktemp ++ local LAST_ERR=/tmp/tmp.karXgnwXeK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VCuDErmfNv ++ cat /tmp/tmp.karXgnwXeK ++ rm /tmp/tmp.VCuDErmfNv /tmp/tmp.karXgnwXeK ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lU4EEt3jA3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pFBPgX1fVS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lU4EEt3jA3 ++ cat /tmp/tmp.pFBPgX1fVS ++ rm /tmp/tmp.lU4EEt3jA3 /tmp/tmp.pFBPgX1fVS ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.IwRA4CiXkx ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.4WRwT5hBA0 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.IwRA4CiXkx +++++ cat /tmp/tmp.4WRwT5hBA0 +++++ rm /tmp/tmp.IwRA4CiXkx /tmp/tmp.4WRwT5hBA0 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1If89Mz4dM +++ mktemp ++ local LAST_ERR=/tmp/tmp.NQQzg8ydqn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1If89Mz4dM ++ cat /tmp/tmp.NQQzg8ydqn ++ rm /tmp/tmp.1If89Mz4dM /tmp/tmp.NQQzg8ydqn ++ return 0 + [[ 3 == \3 ]] + echo + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + check_generation 2 haproxy some-name + local generation=2 + local container=haproxy + local cluster=some-name + local current_generation ++ kubectl_bin get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xKABkQdhfR +++ mktemp ++ local LAST_ERR=/tmp/tmp.GMuiilKPG7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xKABkQdhfR ++ cat /tmp/tmp.GMuiilKPG7 ++ rm /tmp/tmp.xKABkQdhfR /tmp/tmp.GMuiilKPG7 ++ return 0 + current_generation=2 + [[ 2 != \2 ]] + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.y8LRnkpkJw ++ mktemp + local LAST_ERR=/tmp/tmp.PiehQcAbuv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.y8LRnkpkJw secret/my-cluster-secrets patched + cat /tmp/tmp.PiehQcAbuv + rm /tmp/tmp.y8LRnkpkJw /tmp/tmp.PiehQcAbuv + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xgco61mLBU +++ mktemp ++ local LAST_ERR=/tmp/tmp.RLkxl35VLr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xgco61mLBU ++ cat /tmp/tmp.RLkxl35VLr ++ rm /tmp/tmp.xgco61mLBU /tmp/tmp.RLkxl35VLr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.85M550yha8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CLg8dPOvPJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.85M550yha8 ++ cat /tmp/tmp.CLg8dPOvPJ ++ rm /tmp/tmp.85M550yha8 /tmp/tmp.CLg8dPOvPJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mGXNKMdNHt +++ mktemp ++ local LAST_ERR=/tmp/tmp.83OF02nRBt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mGXNKMdNHt ++ cat /tmp/tmp.83OF02nRBt ++ rm /tmp/tmp.mGXNKMdNHt /tmp/tmp.83OF02nRBt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GSUcifXzjh +++ mktemp ++ local LAST_ERR=/tmp/tmp.GMJ9EwJe4n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GSUcifXzjh ++ cat /tmp/tmp.GMJ9EwJe4n ++ rm /tmp/tmp.GSUcifXzjh /tmp/tmp.GMJ9EwJe4n ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jJ2QQbF11b +++ mktemp ++ local LAST_ERR=/tmp/tmp.cnxd793hZY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jJ2QQbF11b ++ cat /tmp/tmp.cnxd793hZY ++ rm /tmp/tmp.jJ2QQbF11b /tmp/tmp.cnxd793hZY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xxmC6IlNLw +++ mktemp ++ local LAST_ERR=/tmp/tmp.SPFYso9e9Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xxmC6IlNLw ++ cat /tmp/tmp.SPFYso9e9Y ++ rm /tmp/tmp.xxmC6IlNLw /tmp/tmp.SPFYso9e9Y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CudYNRx0Xm +++ mktemp ++ local LAST_ERR=/tmp/tmp.woOcSiXXUL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CudYNRx0Xm ++ cat /tmp/tmp.woOcSiXXUL ++ rm /tmp/tmp.CudYNRx0Xm /tmp/tmp.woOcSiXXUL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FKTNf7WxTJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.sC8UHXEbVx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FKTNf7WxTJ ++ cat /tmp/tmp.sC8UHXEbVx ++ rm /tmp/tmp.FKTNf7WxTJ /tmp/tmp.sC8UHXEbVx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x3dPn2swSo +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qx2R9luUfK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.x3dPn2swSo ++ cat /tmp/tmp.Qx2R9luUfK ++ rm /tmp/tmp.x3dPn2swSo /tmp/tmp.Qx2R9luUfK ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NH8h7iACpw +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z57ZYU3bFA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NH8h7iACpw ++ cat /tmp/tmp.Z57ZYU3bFA ++ rm /tmp/tmp.NH8h7iACpw /tmp/tmp.Z57ZYU3bFA ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.K80XYTemfb ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.1VYCb9pkdx +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.K80XYTemfb +++++ cat /tmp/tmp.1VYCb9pkdx +++++ rm /tmp/tmp.K80XYTemfb /tmp/tmp.1VYCb9pkdx +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z0nr2WQTdB +++ mktemp ++ local LAST_ERR=/tmp/tmp.OEGaMCNsNt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Z0nr2WQTdB ++ cat /tmp/tmp.OEGaMCNsNt ++ rm /tmp/tmp.Z0nr2WQTdB /tmp/tmp.OEGaMCNsNt ++ return 0 + [[ 3 == \3 ]] + echo + compare_mysql_cmd select-3 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-3-57.sql ]] + run_mysql 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eTVgutX6w5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tuIGIhIWj0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eTVgutX6w5 ++ cat /tmp/tmp.tuIGIhIWj0 ++ rm /tmp/tmp.eTVgutX6w5 /tmp/tmp.tuIGIhIWj0 ++ return 0 + client_pod=pxc-client-857d976497-fnp6l + wait_pod pxc-client-857d976497-fnp6l + local pod=pxc-client-857d976497-fnp6l + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-fnp6l ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-fnp6l condition met waiting for pod/pxc-client-857d976497-fnp6l to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.miV4rrH5zD/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2232/e2e-tests/users/compare/select-3.sql /tmp/tmp.miV4rrH5zD/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + check_generation 3 haproxy some-name + local generation=3 + local container=haproxy + local cluster=some-name + local current_generation ++ kubectl_bin get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SfLzLQ0BG8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.y2vuFA1U8B ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SfLzLQ0BG8 ++ cat /tmp/tmp.y2vuFA1U8B ++ rm /tmp/tmp.SfLzLQ0BG8 /tmp/tmp.y2vuFA1U8B ++ return 0 + current_generation=3 + [[ 3 != \3 ]] + destroy users-3672 + local namespace=users-3672 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v level=info + grep -v 'the object has been modified' + grep -v 'get backup status: Job.batch' + tee /tmp/tmp.miV4rrH5zD/operator.log ++ get_operator_pod + sort -u ++ local label_prefix=app.kubernetes.io/ + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.hMyOF8dngJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.SWbiYc3iKs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hMyOF8dngJ ++ cat /tmp/tmp.SWbiYc3iKs ++ rm /tmp/tmp.hMyOF8dngJ /tmp/tmp.SWbiYc3iKs ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-58ffd6cd64-hgk7p ++ mktemp + local LAST_OUT=/tmp/tmp.xSvtIsWhR6 ++ mktemp + local LAST_ERR=/tmp/tmp.BeXq555Kto + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-58ffd6cd64-hgk7p + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xSvtIsWhR6 + cat /tmp/tmp.BeXq555Kto + rm /tmp/tmp.xSvtIsWhR6 /tmp/tmp.BeXq555Kto + return 0 2025-11-27T10:35:47.174Z INFO setup Manager starting up {"gitCommit": "31b0cef68f257ed99f383505cbcf3835657757ab", "gitBranch": "PR-2232-31b0cef6", "buildTime": "2025-11-27T08:15:04Z", "goVersion": "go1.25.4", "os": "linux", "arch": "amd64"} 2025-11-27T10:35:47.174Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.31.13-gke.1454000"} 2025-11-27T10:35:47.177Z INFO setup Registering Components. 2025-11-27T10:35:47.950Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2025-11-27T10:35:47.950Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2025-11-27T10:35:47.950Z INFO controller-runtime.metrics Starting metrics server 2025-11-27T10:35:47.950Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2025-11-27T10:35:47.950Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2025-11-27T10:35:47.950Z INFO controller-runtime.webhook Starting webhook server 2025-11-27T10:35:47.950Z INFO setup Starting the Cmd. 2025-11-27T10:35:47.950Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2025-11-27T10:35:47.951Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2025-11-27T10:35:48.051Z INFO attempting to acquire leader lease pxc-operator/08db1feb.percona.com... 2025-11-27T10:35:48.089Z DEBUG events percona-xtradb-cluster-operator-58ffd6cd64-hgk7p_ebf2c2ae-3f29-4702-9632-9e3eb2750bcd became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"575ef1d6-7819-46da-8bde-341d264ac9f9","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1764239748083279009"}, "reason": "LeaderElection"} 2025-11-27T10:35:48.089Z INFO Starting EventSource {"controller": "pxcbackup-controller", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2025-11-27T10:35:48.089Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.PerconaXtraDBCluster"} 2025-11-27T10:35:48.089Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.Secret"} 2025-11-27T10:35:48.089Z INFO Starting EventSource {"controller": "pxcrestore-controller", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2025-11-27T10:35:48.089Z INFO successfully acquired lease pxc-operator/08db1feb.percona.com 2025-11-27T10:35:48.190Z INFO Starting Controller {"controller": "pxcbackup-controller"} 2025-11-27T10:35:48.190Z INFO Starting Controller {"controller": "pxc-controller"} 2025-11-27T10:35:48.190Z INFO Starting Controller {"controller": "pxcrestore-controller"} 2025-11-27T10:35:48.190Z INFO Starting workers {"controller": "pxcbackup-controller", "worker count": 1} 2025-11-27T10:35:48.190Z INFO Starting workers {"controller": "pxc-controller", "worker count": 1} 2025-11-27T10:35:48.190Z INFO Starting workers {"controller": "pxcrestore-controller", "worker count": 1} 2025-11-27T10:36:24.889Z INFO Set CR version {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "version": "1.19.0"} 2025-11-27T10:36:25.061Z INFO User secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "secrets": "my-cluster-secrets"} 2025-11-27T10:36:25.292Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-27T10:36:25.411Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-27T10:36:25.445Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-27T10:36:25.553Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-27T10:36:25.612Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-27T10:36:25.676Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-27T10:36:25.744Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "36e64e81-9a04-4d52-b447-abe93578a2b5", "object": "some-name-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-27T10:36:26.589Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "ab9767a4-d6b3-4d72-aab1-0ec7a7a3b844", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-27T10:36:26.632Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "ab9767a4-d6b3-4d72-aab1-0ec7a7a3b844", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-27T10:37:42.866Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0e443bb0-638e-4b63-9b1a-6f11c6a5037d", "user": "operator"} 2025-11-27T10:37:42.916Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0e443bb0-638e-4b63-9b1a-6f11c6a5037d", "user": "monitor"} 2025-11-27T10:37:42.972Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0e443bb0-638e-4b63-9b1a-6f11c6a5037d"} 2025-11-27T10:37:43.013Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0e443bb0-638e-4b63-9b1a-6f11c6a5037d", "user": "xtrabackup"} 2025-11-27T10:37:43.045Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0e443bb0-638e-4b63-9b1a-6f11c6a5037d"} 2025-11-27T10:37:43.053Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0e443bb0-638e-4b63-9b1a-6f11c6a5037d", "err": "get primary pxc pod: not found"} 2025-11-27T10:37:47.868Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "decad703-3098-4d91-a8fd-9b7dd20c8d39", "err": "get primary pxc pod: not found"} 2025-11-27T10:37:52.999Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "a8aca651-5c1a-4546-8581-6df90d265957", "err": "get primary pxc pod: not found"} 2025-11-27T10:37:58.151Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "fc3e5ea4-abd6-47dc-be53-558d1fc434a8", "err": "get primary pxc pod: not found"} 2025-11-27T10:40:14.265Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c0c2602f-4039-4666-b75a-d10ea0e5ac89", "user": "root"} 2025-11-27T10:40:14.301Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c0c2602f-4039-4666-b75a-d10ea0e5ac89", "user": "replication"} 2025-11-27T10:40:14.351Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c0c2602f-4039-4666-b75a-d10ea0e5ac89", "new version": "5.7.44-48-57"} 2025-11-27T10:40:16.102Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c0c2602f-4039-4666-b75a-d10ea0e5ac89"} 2025-11-27T10:40:21.485Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "53bdccc1-a2e8-4a7d-a4a5-e361c5605e64"} 2025-11-27T10:40:26.232Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "1d362030-311b-4ab7-9ed5-f45888c1fed8"} 2025-11-27T10:40:31.585Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "849b8576-93a9-434a-b928-bb4dae6fa370"} 2025-11-27T10:40:36.902Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "04a82542-7086-42f2-9639-296979f82869"} 2025-11-27T10:40:42.206Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c3f0c8a9-cf8e-42fe-be48-a80e8d6d8a4e"} 2025-11-27T10:40:47.531Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "fafcab8e-aa44-4e6e-a248-44c0366a4635"} 2025-11-27T10:40:52.421Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "a20c78c1-52b6-46f1-8206-8c50ea51f927"} 2025-11-27T10:40:57.893Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "3cb749df-b91b-43a5-ad4a-4f519cedbf88"} 2025-11-27T10:41:03.089Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "540c2a0f-a7e0-4788-947d-52ce476f40da"} 2025-11-27T10:41:08.390Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c738f4fc-9a0a-4d6a-b25f-c138b3f9e9f4"} 2025-11-27T10:41:13.607Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "df56d296-c31e-4702-9a0f-2a309374592b"} 2025-11-27T10:41:18.824Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "2d9dae08-74ed-4d13-969c-4e55a2afd07e"} 2025-11-27T10:41:24.092Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "40eaf1b3-dc2a-4cea-a3fb-f4d015aaa48e"} 2025-11-27T10:41:29.012Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "4db72091-67bf-41fe-81ea-137a80616619"} 2025-11-27T10:41:34.430Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "1f7c54dc-d8a4-470a-a5ce-4e0867083adf"} 2025-11-27T10:41:39.803Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7791958d-82e5-48d4-884a-90216d2ee790"} 2025-11-27T10:41:44.986Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "6895efce-cb9e-4097-ba68-46705983c2d1"} 2025-11-27T10:41:50.153Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "971aae41-dfd9-4cf5-8d8a-0244039e69a3"} 2025-11-27T10:41:51.822Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "79eb5940-78d5-4e12-ac4b-40ca2c685bac", "user": "root"} 2025-11-27T10:41:51.844Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "79eb5940-78d5-4e12-ac4b-40ca2c685bac", "user": "root"} 2025-11-27T10:41:51.863Z INFO MySQL init secret created {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "79eb5940-78d5-4e12-ac4b-40ca2c685bac", "secret": "some-name-mysql-init", "user": "root"} 2025-11-27T10:41:54.318Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "79eb5940-78d5-4e12-ac4b-40ca2c685bac"} 2025-11-27T10:41:54.339Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "79eb5940-78d5-4e12-ac4b-40ca2c685bac", "user": "root"} 2025-11-27T10:41:56.234Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "79eb5940-78d5-4e12-ac4b-40ca2c685bac"} 2025-11-27T10:42:01.316Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "cb6c526f-287b-4afd-9218-d03d7d0c98df"} 2025-11-27T10:42:07.109Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "89c4441d-8d79-40f9-ab59-15a48886d018"} 2025-11-27T10:42:12.343Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "d6576152-6a44-423e-b6c9-4e2fe0d4525b", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:42:12.408Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "d6576152-6a44-423e-b6c9-4e2fe0d4525b", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:42:12.673Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "a5b88241-cc3c-40b4-8799-ecaa23d73e33"} 2025-11-27T10:42:31.829Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "df9280ce-73d3-4ec2-be71-38a205befd6e", "err": "get primary pxc pod: not found"} 2025-11-27T10:42:32.612Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "4fa054a4-a2f4-45de-91da-7790c8b4fe04", "err": "get primary pxc pod: not found"} 2025-11-27T10:42:36.839Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "4fa054a4-a2f4-45de-91da-7790c8b4fe04", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:42:38.278Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f72f1a7b-924e-4c9f-b8f1-224587efa13d", "user": "proxyadmin"} 2025-11-27T10:42:38.278Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f72f1a7b-924e-4c9f-b8f1-224587efa13d", "user": "proxyadmin"} 2025-11-27T10:42:38.339Z INFO Proxy user updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f72f1a7b-924e-4c9f-b8f1-224587efa13d", "user": "proxyadmin"} 2025-11-27T10:42:38.363Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f72f1a7b-924e-4c9f-b8f1-224587efa13d", "user": "proxyadmin"} 2025-11-27T10:42:38.363Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f72f1a7b-924e-4c9f-b8f1-224587efa13d", "last-applied-secret": "5d0152d43aa8f626994e4734d78ab239a0f25610406c1ae08be1258a4104fbbe"} 2025-11-27T10:42:38.369Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f72f1a7b-924e-4c9f-b8f1-224587efa13d", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:42:38.421Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f72f1a7b-924e-4c9f-b8f1-224587efa13d", "err": "get primary pxc pod: not found"} 2025-11-27T10:42:39.906Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "fd32af86-9737-43dc-b9c5-b4f7efbf5a14", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / \nSyncing user accounts from PXC(some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local:3306) to ProxySQL\nAdding user to ProxySQL: replication\n / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:797) : Failed to load user list from ProxySQL database. \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:718) : Failed to add the user (replication) from PXC to ProxySQL database. \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / \nSyncing user accounts from PXC(some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local:3306) to ProxySQL\nAdding user to ProxySQL: replication\n / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:797) : Failed to load user list from ProxySQL database. \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:718) : Failed to add the user (replication) from PXC to ProxySQL database. \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:42:44.569Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "15ee9404-205b-42f9-ac23-a83f914e707a", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:43:12.917Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5c1051b3-4e18-4d04-825b-b4557841ffa7"} 2025-11-27T10:43:18.008Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "43fb9bfa-d9a9-401b-91a9-11c00b54acd9"} 2025-11-27T10:43:21.004Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "32531c58-5015-4943-bd5b-63d377c470d9", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:43:21.082Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "32531c58-5015-4943-bd5b-63d377c470d9", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:43:23.416Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c9789a6d-16af-42f5-9163-eba84887c3b1", "user": "xtrabackup"} 2025-11-27T10:43:23.427Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c9789a6d-16af-42f5-9163-eba84887c3b1", "user": "xtrabackup"} 2025-11-27T10:43:23.448Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c9789a6d-16af-42f5-9163-eba84887c3b1", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2025-11-27T10:43:23.473Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c9789a6d-16af-42f5-9163-eba84887c3b1", "user": "xtrabackup"} 2025-11-27T10:43:23.473Z INFO PXC pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c9789a6d-16af-42f5-9163-eba84887c3b1", "last-applied-secret": "2eeafa734e134261f95958f0b91868bbfc622722ec539ea902872291e850a1d6"} 2025-11-27T10:43:23.477Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c9789a6d-16af-42f5-9163-eba84887c3b1", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:43:26.118Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c9789a6d-16af-42f5-9163-eba84887c3b1"} 2025-11-27T10:45:16.580Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f2b28abe-33c5-456d-92ca-7a1a3b90aeac", "err": "failed to connect to pod some-name-pxc-0: dial tcp 10.119.40.47:33062: connect: connection refused"} 2025-11-27T10:45:21.745Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0cce862b-9650-40e0-a112-b377c096aee7", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:45:26.880Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "fcfa72ed-b04d-4d14-9492-2b69368bb204", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:45:32.020Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "e28f23ce-4ce7-459d-9c62-399d4e674b15", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:45:37.215Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "d3555057-8482-4556-9a3d-07c8b70385f5", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:45:42.362Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "8edcb4c8-0574-484b-8d26-a60e1f9fcd11", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:45:47.562Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "42e3e7f6-6eb0-491a-b2e2-be3b51b7cad9", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:45:52.693Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "4fed6a76-af1c-4166-8502-27597d440cb3", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:46:00.434Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "08dfe6cf-948c-4856-83e1-f97fb92fa03f"} 2025-11-27T10:46:05.357Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "deda166a-2a2d-4b5c-9297-ad0d0bba9538"} 2025-11-27T10:46:06.335Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "user": "monitor"} 2025-11-27T10:46:06.347Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "user": "monitor"} 2025-11-27T10:46:06.377Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "secret": "some-name-mysql-init", "user": "monitor"} 2025-11-27T10:46:06.396Z INFO Proxy user updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "user": "monitor"} 2025-11-27T10:46:06.418Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "user": "monitor"} 2025-11-27T10:46:06.418Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "last-applied-secret": "ff1456fdbc5f20df9bcdefa9705214611570892a1a58c611c725bc14da1393ca"} 2025-11-27T10:46:06.422Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:46:08.991Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "5bc2e75f-cc7b-4e74-b37b-9a408b88b957", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:46:55.199Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "727d2ca9-a96f-4303-81f1-1d32f2ee277c"} 2025-11-27T10:47:00.111Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "b627276b-920f-4250-8616-a5e110473ab8"} 2025-11-27T10:47:05.325Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "4329e101-5688-4a58-9ef3-8b50cc9504dc"} 2025-11-27T10:47:10.587Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "53897c64-4767-4bbb-9813-4a1820e6bd8b"} 2025-11-27T10:47:11.284Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7d621765-0730-4746-881a-0662f1030223", "user": "operator"} 2025-11-27T10:47:11.296Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7d621765-0730-4746-881a-0662f1030223", "user": "operator"} 2025-11-27T10:47:11.317Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7d621765-0730-4746-881a-0662f1030223", "secret": "some-name-mysql-init", "user": "operator"} 2025-11-27T10:47:11.338Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7d621765-0730-4746-881a-0662f1030223", "user": "operator"} 2025-11-27T10:47:11.338Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7d621765-0730-4746-881a-0662f1030223", "last-applied-secret": "76985cf1a9c027294ebef24d382b78b998ade963d34470733949bdb4a17deb86"} 2025-11-27T10:47:11.342Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7d621765-0730-4746-881a-0662f1030223", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:47:15.062Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7d621765-0730-4746-881a-0662f1030223", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:47:40.276Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "d5901129-d637-4621-9783-a4ab143339fb"} 2025-11-27T10:47:44.843Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "736285e2-f65f-42dc-8be2-a99012be22a6"} 2025-11-27T10:47:50.160Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "353ee289-feff-434d-a283-592cb94e4eec"} 2025-11-27T10:47:55.212Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "e44e902b-04d2-4b1d-a221-5956ce74bd9d"} 2025-11-27T10:47:57.286Z INFO Created user secrets {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "secrets": "my-cluster-secrets-2"} 2025-11-27T10:47:57.286Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "root"} 2025-11-27T10:47:57.301Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "root"} 2025-11-27T10:47:57.322Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "secret": "some-name-mysql-init", "user": "root"} 2025-11-27T10:47:59.862Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b"} 2025-11-27T10:47:59.883Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "root"} 2025-11-27T10:47:59.883Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "operator"} 2025-11-27T10:47:59.894Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "operator"} 2025-11-27T10:47:59.913Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "secret": "some-name-mysql-init", "user": "operator"} 2025-11-27T10:47:59.932Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "operator"} 2025-11-27T10:47:59.932Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "monitor"} 2025-11-27T10:47:59.943Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "monitor"} 2025-11-27T10:47:59.969Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "secret": "some-name-mysql-init", "user": "monitor"} 2025-11-27T10:47:59.987Z INFO Proxy user updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "monitor"} 2025-11-27T10:48:00.013Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "monitor"} 2025-11-27T10:48:00.013Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "xtrabackup"} 2025-11-27T10:48:00.023Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "xtrabackup"} 2025-11-27T10:48:00.058Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2025-11-27T10:48:00.089Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "xtrabackup"} 2025-11-27T10:48:00.089Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "replication"} 2025-11-27T10:48:00.103Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "replication"} 2025-11-27T10:48:00.127Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "secret": "some-name-mysql-init", "user": "replication"} 2025-11-27T10:48:00.155Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "replication"} 2025-11-27T10:48:00.155Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "proxyadmin"} 2025-11-27T10:48:00.174Z INFO Proxy user updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "proxyadmin"} 2025-11-27T10:48:00.193Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "user": "proxyadmin"} 2025-11-27T10:48:00.193Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "last-applied-secret": "89db85768a266e79c2a221e266ec28fed7ad6281b8798ebe62eb8d9dfd4e9029"} 2025-11-27T10:48:00.193Z INFO PXC pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "last-applied-secret": "89db85768a266e79c2a221e266ec28fed7ad6281b8798ebe62eb8d9dfd4e9029"} 2025-11-27T10:48:00.196Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:48:00.252Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:48:02.272Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7f05d4f7-fe21-453c-ae12-3f6d0077002b", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:49:47.976Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "25248ec9-768b-4e21-9a09-256775e16a45", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:50:13.724Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "d983967f-8f54-47ae-9e65-8e29126f876e", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:50:21.491Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "2ab55d79-9586-42da-af60-0c09681d19cf"} 2025-11-27T10:50:26.551Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "99ead664-ffb7-46a5-ae28-9773fe62ffe6"} 2025-11-27T10:50:27.760Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "671c5535-e43d-4289-9bc6-580e4c42a574", "user": "operator"} 2025-11-27T10:50:27.772Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "671c5535-e43d-4289-9bc6-580e4c42a574", "user": "operator"} 2025-11-27T10:50:27.793Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "671c5535-e43d-4289-9bc6-580e4c42a574", "secret": "some-name-mysql-init", "user": "operator"} 2025-11-27T10:50:27.813Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "671c5535-e43d-4289-9bc6-580e4c42a574", "user": "operator"} 2025-11-27T10:50:27.813Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "671c5535-e43d-4289-9bc6-580e4c42a574", "last-applied-secret": "787123194f9bf5127f28a13c02a85e4a52b45f424a48ca9561efdcc9d6d780e8"} 2025-11-27T10:50:27.817Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "671c5535-e43d-4289-9bc6-580e4c42a574", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:50:31.380Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "671c5535-e43d-4289-9bc6-580e4c42a574", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-3672.svc.c' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:51:21.875Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "554b07e5-7d51-4064-9592-39ca0f2f96bd"} 2025-11-27T10:51:26.373Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "f4167d4e-551b-4fbd-9cf9-dc1c9f0f9049"} 2025-11-27T10:51:31.488Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "b28b207a-95a9-4bbd-8d7a-cd356d6b4db3"} 2025-11-27T10:51:36.881Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c1174967-6280-43d2-9062-5d6bca8d50be"} 2025-11-27T10:51:42.174Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "749d2d5a-f80a-4903-a943-6247a9c00425"} 2025-11-27T10:51:47.164Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "13680457-0c0c-4ede-ae09-d1ef490eb116"} 2025-11-27T10:51:52.463Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "7aa63dcd-c173-4351-a118-70ee3c0ad896"} 2025-11-27T10:51:58.489Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "281a0f0a-ef48-4cbc-a362-bf6033b8ffeb"} 2025-11-27T10:52:03.253Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "031b8ddc-00e5-4656-921b-a9fcf6c44bf2"} 2025-11-27T10:52:08.538Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "2b255dde-48f9-4f85-83fb-7c1314f7011f"} 2025-11-27T10:52:13.750Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "bb0277b7-bcb7-4274-99e2-fc470c5db8ea"} 2025-11-27T10:52:18.950Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "b61d2f48-7353-4c1c-ae9e-2e600fb54d18"} 2025-11-27T10:52:23.957Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "d34e0cdb-cbe6-4180-8e9f-59412f36c136"} 2025-11-27T10:52:29.242Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0cb05818-e29a-440d-bb7e-7db62a8bc676"} 2025-11-27T10:52:34.732Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "93a9ce2c-0c90-483a-95f7-e3f0c84c4ead"} 2025-11-27T10:52:40.067Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "ef923b15-cab3-42f4-bc87-2920249fca5d"} 2025-11-27T10:52:40.752Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "root"} 2025-11-27T10:52:40.768Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "root"} 2025-11-27T10:52:40.787Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "secret": "some-name-mysql-init", "user": "root"} 2025-11-27T10:52:43.395Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0"} 2025-11-27T10:52:43.416Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "root"} 2025-11-27T10:52:43.416Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "monitor"} 2025-11-27T10:52:43.426Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "monitor"} 2025-11-27T10:52:43.446Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "secret": "some-name-mysql-init", "user": "monitor"} 2025-11-27T10:52:43.465Z INFO Proxy user updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "monitor"} 2025-11-27T10:52:43.486Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "monitor"} 2025-11-27T10:52:43.486Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "xtrabackup"} 2025-11-27T10:52:43.498Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "xtrabackup"} 2025-11-27T10:52:43.516Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2025-11-27T10:52:43.535Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "xtrabackup"} 2025-11-27T10:52:43.535Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "proxyadmin"} 2025-11-27T10:52:43.555Z INFO Proxy user updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "proxyadmin"} 2025-11-27T10:52:43.576Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "user": "proxyadmin"} 2025-11-27T10:52:43.576Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "last-applied-secret": "352ec183c1b6b8ef93cea415a5bd6f047e75886a57fdd8d864e7cd06f4528aed"} 2025-11-27T10:52:43.576Z INFO PXC pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "last-applied-secret": "352ec183c1b6b8ef93cea415a5bd6f047e75886a57fdd8d864e7cd06f4528aed"} 2025-11-27T10:52:43.579Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:52:43.668Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:52:45.333Z ERROR sync users {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c4d7e467-5b56-476c-9e25-98bf573b7ec0", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'127.0.0.1' (using password: YES)\nERROR (line:339) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at localhost:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:854\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-27T10:54:36.540Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "b94dd65a-0423-416e-a300-a25bbf2edc5c", "err": "failed to connect to pod some-name-pxc-0: dial tcp 10.119.40.54:33062: connect: connection refused"} 2025-11-27T10:54:41.726Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "c530f059-c6a1-4740-8bba-53027e1268c9", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:54:46.852Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "6845c8eb-c173-49c4-ab0b-9d6e3ca06497", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:54:51.995Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "237d2fd5-10e3-4854-a3f0-1723d8ff1d26", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:54:57.119Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "d4f777cd-dddb-44bd-8df8-4bf5bfbdbcd3", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:55:02.278Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "90eb1b04-3c98-4e50-9e3b-663f00c7b60d", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:55:07.456Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "a61275cc-c33f-4581-9f56-74ceaeecf175", "primary name": "some-name-pxc-0.some-name-pxc.users-3672.svc.cluster.local"} 2025-11-27T10:55:15.321Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "3d5a203b-e227-4615-8d0f-2d4490a148fa"} 2025-11-27T10:55:18.041Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "root"} 2025-11-27T10:55:18.057Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "root"} 2025-11-27T10:55:18.104Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "secret": "some-name-mysql-init", "user": "root"} 2025-11-27T10:55:18.130Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "root"} 2025-11-27T10:55:18.130Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "operator"} 2025-11-27T10:55:18.144Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "operator"} 2025-11-27T10:55:18.205Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "secret": "some-name-mysql-init", "user": "operator"} 2025-11-27T10:55:18.243Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "operator"} 2025-11-27T10:55:18.244Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "monitor"} 2025-11-27T10:55:18.254Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "monitor"} 2025-11-27T10:55:18.284Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "secret": "some-name-mysql-init", "user": "monitor"} 2025-11-27T10:55:18.305Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "monitor"} 2025-11-27T10:55:18.305Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "xtrabackup"} 2025-11-27T10:55:18.316Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "xtrabackup"} 2025-11-27T10:55:18.335Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2025-11-27T10:55:18.355Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "xtrabackup"} 2025-11-27T10:55:18.355Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "replication"} 2025-11-27T10:55:18.365Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "replication"} 2025-11-27T10:55:18.384Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "secret": "some-name-mysql-init", "user": "replication"} 2025-11-27T10:55:18.406Z INFO HAProxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "last-applied-secret": "76985cf1a9c027294ebef24d382b78b998ade963d34470733949bdb4a17deb86"} 2025-11-27T10:55:18.406Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "user": "replication"} 2025-11-27T10:55:18.406Z INFO PXC pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "last-applied-secret": "76985cf1a9c027294ebef24d382b78b998ade963d34470733949bdb4a17deb86"} 2025-11-27T10:55:18.407Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:55:18.456Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-27T10:55:18.515Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:55:18.620Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2025-11-27T10:55:18.711Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-27T10:55:18.835Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "object": "some-name-haproxy-replicas", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-27T10:55:22.451Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "35ed1c3b-f938-4c21-b7f9-e928f05064fd", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 34.118.238.11:3306: connect: connection refused"} 2025-11-27T10:55:23.232Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "dc456535-6231-48ca-9841-54964cf1118b", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-27T10:56:52.138Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "a22d1af4-e925-415c-a667-44fb49449a14", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-3672 on 34.118.224.10:53: no such host"} 2025-11-27T10:56:52.457Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "0de93da5-99a7-47e0-b969-a046c926b551", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-3672 on 34.118.224.10:53: no such host"} 2025-11-27T10:57:40.073Z INFO Password changed, updating user {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "27b56d40-9e8a-4fdd-b1e2-e6b87de7d285", "user": "monitor"} 2025-11-27T10:57:40.084Z INFO User password updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "27b56d40-9e8a-4fdd-b1e2-e6b87de7d285", "user": "monitor"} 2025-11-27T10:57:40.105Z INFO MySQL init secret updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "27b56d40-9e8a-4fdd-b1e2-e6b87de7d285", "secret": "some-name-mysql-init", "user": "monitor"} 2025-11-27T10:57:40.127Z INFO HAProxy pods will be restarted {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "27b56d40-9e8a-4fdd-b1e2-e6b87de7d285", "last-applied-secret": "991826bff863e86ff93e85044f694dc7ba0d45b478b78bf25e528bb1e9f26246"} 2025-11-27T10:57:40.127Z INFO Internal secrets updated {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "27b56d40-9e8a-4fdd-b1e2-e6b87de7d285", "user": "monitor"} 2025-11-27T10:57:40.130Z DEBUG Updating object {"controller": "pxc-controller", "namespace": "users-3672", "name": "some-name", "reconcileID": "27b56d40-9e8a-4fdd-b1e2-e6b87de7d285", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:856 [mysql] 2025/11/27 10:56:58 packets.go:58 unexpected EOF [mysql] 2025/11/27 10:57:12 packets.go:58 unexpected EOF -  }, -  { -  }, -  { -  }, -  }, +  }, -  "352ec183c1b6b8ef93cea415a5bd6f047e75886a57fdd8d864e7cd06f4528aed", +  "352ec183c1b6b8ef93cea415a5bd6f047e75886a57fdd8d864e7cd06f4528aed", +  "66e79c2a221e266ec28fed7ad6281b8798ebe62eb8d9dfd4e9029", +  "6fdbc5f20df9bcdefa9705214611570892a1a58c611c725bc14da1393ca", -  "76985cf1a9c027294ebef24d382b78b998ade963d34470733949bdb4a17deb8", -  "76985cf1a9c027294ebef24d382b78b998ade963d34470733949bdb4a17deb86", +  "76985cf1a9c027294ebef24d382b78b998ade963d34470733949bdb4a17deb86", -  "787123194f9bf5127f28a13c02a85e4a52b45f424a48ca9561efdcc9d6d780e8", +  "787123194f9bf5127f28a13c02a85e4a52b45f424a48ca9561efdcc9d6d780e8", +  "89db85768a", -  "89db85768a266e79c2a221e266ec28fed7ad6281b8798ebe62eb8d9dfd4e9029", +  "89db85768a266e79c2a221e266ec28fed7ad6281b8798ebe62eb8d9dfd4e9029", +  "991826bff863e86ff93e85044f694dc7ba0d45b478b78bf25e528bb1e9f2624", -  Annotations: map[string]string{ +  Annotations: map[string]string{ +  APIVersion: "", -  APIVersion: "apps/v1", -  APIVersion: "apps/v1", -  APIVersion: "v1", -  Args: []string{"logrotate"}, +  AvailableReplicas: 0, -  AvailableReplicas: 2, -  AvailableReplicas: 3, -  CollisionCount: &0, +  CollisionCount: nil, +  CreationTimestamp: v1.Time{}, -  CreationTimestamp: v1.Time{Time: s"2025-11-27 10:36:25 +0000 UTC"}, -  CreationTimestamp: v1.Time{Time: s"2025-11-27 10:55:18 +0000 UTC"}, +  CurrentReplicas: 0, -  CurrentReplicas: 2, -  CurrentReplicas: 3, +  CurrentRevision: "", -  CurrentRevision: "some-name-haproxy-5b567bf96c", -  CurrentRevision: "some-name-proxysql-566898c866", -  CurrentRevision: "some-name-proxysql-5868b47f6b", -  CurrentRevision: "some-name-proxysql-65f5ccc684", -  CurrentRevision: "some-name-proxysql-746b976c8d", -  CurrentRevision: "some-name-proxysql-7965c48b57", -  CurrentRevision: "some-name-proxysql-f874d7649", -  CurrentRevision: "some-name-pxc-6b45b7d479", -  CurrentRevision: "some-name-pxc-766887bc84", -  CurrentRevision: "some-name-pxc-77d4f8d758", -  CurrentRevision: "some-name-pxc-7d5d5c48bb", -  "d0152d43aa8f626994e4734d78ab239a0f25610406c1ae08be1258a4104fbbe", -  DefaultMode: &420, -  DefaultMode: &420, +  DefaultMode: nil, +  DefaultMode: nil, +  DeprecatedServiceAccount: "", -  DeprecatedServiceAccount: "default", +  DNSPolicy: "", -  DNSPolicy: "ClusterFirst", -  "eeafa734e134261f95958f0b91868bbfc622722ec539ea902872291e850a1d6", -  EnvFrom: []v1.EnvFromSource{{SecretRef: s"&SecretEnvSource{LocalObjectRefe"...}}, -  Env: []v1.EnvVar{ +  "ff14", -  "ff1456fdbc5f20df9bcdefa9705214611570892a1a58c611c725bc14da1393ca", -  FieldsType: "FieldsV1", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., +  Generation: 0, -  Generation: 1, -  Generation: 2, -  Generation: 3, -  Generation: 4, -  Generation: 5, -  Generation: 6, -  Generation: 7, -  Generation: 8, -  Image: "perconalab/percona-xtradb-cluster-operator:main-logcollector", -  Image: "perconalab/percona-xtradb-cluster-operator:main-logcollector", -  ImagePullPolicy: "Always", +  "last-applied-secret": "2eeafa734e134261f95958f0b91868bbfc622722ec539ea902872291e850a1d6", +  "last-applied-secret": "5d0152d43aa8f626994e4734d78ab239a0f25610406c1ae08be1258a4104fbbe", +  "last-applied-secret": "76985cf1a9c027294ebef24d382b78b998ade963d34470733949bdb4a17deb86", +  ManagedFields: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  Manager: "kube-controller-manager", -  Manager: "percona-xtradb-cluster-operator", -  {Name: "IS_LOGCOLLECTOR", Value: "yes"}, -  {Name: "LOG_DATA_DIR", Value: "/var/lib/mysql"}, -  Name: "logrotate", -  Name: "logs", -  {Name: "MONITOR_PASSWORD", ValueFrom: s"&EnvVarSource{FieldRef:nil,Resou"...}, -  {Name: "POD_NAMESPASE", ValueFrom: s"&EnvVarSource{FieldRef:&ObjectFi"...}, -  {Name: "POD_NAME", ValueFrom: s"&EnvVarSource{FieldRef:&ObjectFi"...}, -  {Name: "SERVICE_TYPE", Value: "mysql"}, +  ObservedGeneration: 0, -  ObservedGeneration: 1, -  ObservedGeneration: 2, -  ObservedGeneration: 3, -  ObservedGeneration: 4, -  ObservedGeneration: 5, -  ObservedGeneration: 6, -  ObservedGeneration: 7, -  ObservedGeneration: 8, -  Operation: "Update", -  Operation: "Update", -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMzUyZWMxODNjMWI2YjhlZjkzY2VhNDE1YTViZDZmMDQ3ZTc1ODg2YTU3ZmRkOGQ4NjRlN2NkMDZmNDUyOGFlZCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNWQwMTUyZDQzYWE4ZjYyNjk5NGU0NzM0ZDc4YWIyMzlhMGYyNTYxMDQwNmMxYWUwOGJlMTI1OGE0MTA0ZmJiZSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzg3MTIzMTk0ZjliZjUxMjdmMjhhMTNjMDJhODVlNGE1MmI0NWY0MjRhNDhjYTk1NjFlZmRjYzlkNmQ3ODBlOCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzg3MTIzMTk0ZjliZjUxMjdmMjhhMTNjMDJhODVlNGE1MmI0NWY0MjRhNDhjYTk1NjFlZmRjYzlkNmQ3ODBlOCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzY5ODVjZjFhOWMwMjcyOTRlYmVmMjRkMzgyYjc4Yjk5OGFkZTk2M2QzNDQ3MDczMzk0OWJkYjRhMTdkZWI4NiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzY5ODVjZjFhOWMwMjcyOTRlYmVmMjRkMzgyYjc4Yjk5OGFkZTk2M2QzNDQ3MDczMzk0OWJkYjRhMTdkZWI4NiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiODlkYjg1NzY4YTI2NmU3OWMyYTIyMWUyNjZlYzI4ZmVkN2FkNjI4MWI4Nzk4ZWJlNjJlYjhkOWRmZDRlOTAyOSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiODlkYjg1NzY4YTI2NmU3OWMyYTIyMWUyNjZlYzI4ZmVkN2FkNjI4MWI4Nzk4ZWJlNjJlYjhkOWRmZDRlOTAyOSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZmYxNDU2ZmRiYzVmMjBkZjliY2RlZmE5NzA1MjE0NjExNTcwODkyYTFhNThjNjExYzcyNWJjMTRkYTEzOTNjYSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZmYxNDU2ZmRiYzVmMjBkZjliY2RlZmE5NzA1MjE0NjExNTcwODkyYTFhNThjNjExYzcyNWJjMTRkYTEzOTNjYSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6ImhhcHJveHkiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzY5ODVjZjFhOWMwMjcyOTRlYmVmMjRkMzgyYjc4Yjk5OGFkZTk2M2QzNDQ3MDczMzk0OWJkYjRhMTdkZWI4NiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6ImhhcHJveHkiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzY5ODVjZjFhOWMwMjcyOTRlYmVmMjRkMzgyYjc4Yjk5OGFkZTk2M2QzNDQ3MDczMzk0OWJkYjRhMTdkZWI4NiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6ImhhcHJveHkiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiOTkxODI2YmZmODYzZTg2ZmY5M2U4NTA0NGY2OTRkYzdiYTBkNDViNDc4Yjc4YmYyNWU1MjhiYjFlOWYyNjI0NiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoiaGFwcm94eSIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoic29tZS1uYW1lIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6ImhhcHJveHkiLCJwZXJjb25hLmNvbS9jb25maWd1cmF0aW9uLWhhc2giOiJkNDFkOGNkOThmMDBiMjA0ZTk4MDA5OThlY2Y4NDI3ZSJ9fSwic3BlYyI6eyJ2b2x1bWVzIjpbeyJuYW1lIjoiaGFwcm94eS1jdXN0b20iLCJjb25maWdNYXAi"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMmVlYWZhNzM0ZTEzNDI2MWY5NTk1OGYwYjkxODY4YmJmYzYyMjcyMmVjNTM5ZWE5MDI4NzIyOTFlODUwYTFkNiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMmVlYWZhNzM0ZTEzNDI2MWY5NTk1OGYwYjkxODY4YmJmYzYyMjcyMmVjNTM5ZWE5MDI4NzIyOTFlODUwYTFkNiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMzUyZWMxODNjMWI2YjhlZjkzY2VhNDE1YTViZDZmMDQ3ZTc1ODg2YTU3ZmRkOGQ4NjRlN2NkMDZmNDUyOGFlZCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMzUyZWMxODNjMWI2YjhlZjkzY2VhNDE1YTViZDZmMDQ3ZTc1ODg2YTU3ZmRkOGQ4NjRlN2NkMDZmNDUyOGFlZCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMzUyZWMxODNjMWI2YjhlZjkzY2VhNDE1YTViZDZmMDQ3ZTc1ODg2YTU3ZmRkOGQ4NjRlN2NkMDZmNDUyOGFlZCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJjZmQ2YTUyMzk4MjY4MTczYjUxZDdjZGIxMzMxYzA5YSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiNzUxZTU5YTFjMWQ5NGFjNjdmOWQ2YmI2ZTc1NmFjZGMifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InRtcCIsImVtcHR5RGlyIjp7fX0seyJuYW1lIjoiY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJzb21lLW5hbWUtcHhjIiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJzc2wtaW50ZXJuYWwiLCJzZWNyZXQiOnsic2VjcmV0TmFtZSI6InNvbWUtbmFtZS1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoic29tZS1uYW1lLXNzbCIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6ImF1dG8tY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJhdXRvLXNvbWUtbmFtZS1weGMiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InZhdWx0LWtleXJpbmctc2VjcmV0Iiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJzb21lLW5hbWUtdmF1bHQiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6Im15c3FsLXVzZXJzLXNlY3JldC1maWxlIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJpbnRlcm5hbC1zb21lLW5hbWUiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJteXNxbC1pbml0LWZpbGUiLCJzZWNyZXQiOnsic2VjcmV0TmFtZSI6InNvbWUtbmFtZS1teXNxbC1pbml0Iiwib3B0aW9uYWwiOnRydWV9fV0sImluaXRDb250YWluZXJzIjpbeyJuYW1lIjoicHhjLWluaXQiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjpQUi0yMjMyLTMxYjBjZWY2IiwiY29tbWFuZCI6WyIvcHhjLWluaXQtZW50cnlwb2ludC5zaCJdLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjUwbSIsIm1lbW9yeSI6IjUwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoiZGF0YWRpciIsIm1vdW50UGF0aCI6Ii92YXIvbGliL215c3FsIn1dLCJpbWFnZVB1bGxQb2xpY3kiOiJBbHdheXMifV0sImNvbnRhaW5lcnMiOlt7Im5hbWUiOiJsb2dzIiwiaW1hZ2UiOiJwZXJjb25hbGFiL3BlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3I6bWFpbi1sb2djb2xsZWN0b3IiLCJlbnZGcm9tIjpbeyJzZWNyZXRSZWYiOnsibmFtZSI6InNvbWUtbmFtZS1sb2ctY29sbGVjdG9yIiwib3B0aW9uYWwiOnRydWV9fV0sImVudiI6W3sibmFtZSI6IkxPR19EQVRBX0RJUiIsInZhbHVlIjoiL3Zhci9saWIvbXlzcWwifSx7Im5hbWUiOiJQT0RfTkFNRVNQQVNFIiwidmFsdWVGcm9tIjp7ImZpZWxkUmVmIjp7ImZpZWxkUGF0aCI6Im1ldGFkYXRhLm5hbWVzcGFjZSJ9fX0seyJuYW1lIjoiUE9EX05BTUUiLCJ2YWx1ZUZyb20iOnsiZmllbGRSZWYiOnsiZmllbGRQYXRoIjoibWV0YWRhdGEubmFtZSJ9fX1dLCJyZXNvdXJjZXMiOnt9LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJkYXRhZGlyIiwibW91bnRQYXRoIjoiL3Zhci9saWIvbXlzcWwifV0sImltYWdlUHVsbFBvbGljeSI6IkFsd2F5cyJ9LHsibmFtZSI6ImxvZ3JvdGF0ZSIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOm1haW4tbG9nY29sbGVjdG9yIiwiYXJncyI6WyJsb2dyb3RhdGUiXSwiZW52IjpbeyJuYW1lIjoiU0VSVklDRV9UWVBFIiwidmFsdWUiOiJteXNxbCJ9LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1zb21lLW5hbWUiLCJrZXkiOiJtb25pdG9yIn19fV0sInJlc291cmNlcyI6e30sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImRhdGFkaXIiLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHhjIiwiaW1hZ2UiOiJwZXJjb25hbGFiL3BlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3I6bWFpbi1weGM1LjciLCJjb21tYW5kIjpbIi92YXIvbGliL215c3FsL3B4Yy1lbnRyeXBvaW50LnNoIl0sImFyZ3MiOlsibXlzcWxkIl0sInBvcnRzIjpbeyJuYW1lIjoibXlzcWwiLCJjb250YWluZXJQb3J0IjozMzA2fSx7Im5hbWUiOiJzc3QiLCJjb250YWluZXJQb3J0Ijo0NDQ0fSx7Im5hbWUiOiJ3cml0ZS1zZXQiLCJjb250YWluZXJQb3J0Ijo0NTY3fSx7Im5hbWUiOiJpc3QiLCJjb250YWluZXJQb3J0Ijo0NTY4fSx7Im5hbWUiOiJteXNxbC1hZG1pbiIsImNvbnRhaW5lclBvcnQiOjMzMDYyfSx7Im5h"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMzUyZWMxODNjMWI2YjhlZjkzY2VhNDE1YTViZDZmMDQ3ZTc1ODg2YTU3ZmRkOGQ4NjRlN2NkMDZmNDUyOGFlZCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJjZmQ2YTUyMzk4MjY4MTczYjUxZDdjZGIxMzMxYzA5YSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiNzUxZTU5YTFjMWQ5NGFjNjdmOWQ2YmI2ZTc1NmFjZGMifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InRtcCIsImVtcHR5RGlyIjp7fX0seyJuYW1lIjoiY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJzb21lLW5hbWUtcHhjIiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJzc2wtaW50ZXJuYWwiLCJzZWNyZXQiOnsic2VjcmV0TmFtZSI6InNvbWUtbmFtZS1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoic29tZS1uYW1lLXNzbCIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6ImF1dG8tY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJhdXRvLXNvbWUtbmFtZS1weGMiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InZhdWx0LWtleXJpbmctc2VjcmV0Iiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJzb21lLW5hbWUtdmF1bHQiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6Im15c3FsLXVzZXJzLXNlY3JldC1maWxlIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJpbnRlcm5hbC1zb21lLW5hbWUiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJteXNxbC1pbml0LWZpbGUiLCJzZWNyZXQiOnsic2VjcmV0TmFtZSI6InNvbWUtbmFtZS1teXNxbC1pbml0Iiwib3B0aW9uYWwiOnRydWV9fV0sImluaXRDb250YWluZXJzIjpbeyJuYW1lIjoicHhjLWluaXQiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjpQUi0yMjMyLTMxYjBjZWY2IiwiY29tbWFuZCI6WyIvcHhjLWluaXQtZW50cnlwb2ludC5zaCJdLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjUwbSIsIm1lbW9yeSI6IjUwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoiZGF0YWRpciIsIm1vdW50UGF0aCI6Ii92YXIvbGliL215c3FsIn1dLCJpbWFnZVB1bGxQb2xpY3kiOiJBbHdheXMifV0sImNvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXB4YzUuNyIsImNvbW1hbmQiOlsiL3Zhci9saWIvbXlzcWwvcHhjLWVudHJ5cG9pbnQuc2giXSwiYXJncyI6WyJteXNxbGQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InNzdCIsImNvbnRhaW5lclBvcnQiOjQ0NDR9LHsibmFtZSI6IndyaXRlLXNldCIsImNvbnRhaW5lclBvcnQiOjQ1Njd9LHsibmFtZSI6ImlzdCIsImNvbnRhaW5lclBvcnQiOjQ1Njh9LHsibmFtZSI6Im15c3FsLWFkbWluIiwiY29udGFpbmVyUG9ydCI6MzMwNjJ9LHsibmFtZSI6Im15c3FseCIsImNvbnRhaW5lclBvcnQiOjMzMDYwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoic29tZS1uYW1lLWVudi12YXJzLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoic29tZS1uYW1lLXB4Yy11bnJlYWR5In0seyJuYW1lIjoiTU9OSVRPUl9IT1NUIiwidmFsdWUiOiIlIn0seyJuYW1lIjoiTVlTUUxfUk9PVF9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXNvbWUtbmFtZSIsImtleSI6InJvb3QifX19LHsibmFtZSI6IlhUUkFCQUNLVVBfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1zb21lLW5hbWUiLCJrZXkiOiJ4dHJhYmFja3VwIn19fSx7Im5hbWUiOiJNT05JVE9SX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtc29tZS1uYW1lIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiQ0xVU1RFUl9IQVNIIiwidmFsdWUiOiIxODY3MjI0In0seyJuYW1lIjoiT1BFUkFUT1JfQURNSU5fUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1zb21lLW5hbWUiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiTElWRU5FU1NfQ0hFQ0tfVElNRU9VVCIsInZhbHVlIjoiNSJ9LHsibmFtZSI6IlJFQURJTkVTU19DSEVDS19USU1FT1VUIiwidmFsdWUiOiIxNSJ9LHsibmFtZSI6IkRFRkFVTFRfQVVUSEVOVElDQVRJT05fUExVR0lOIiwidmFsdWUiOiJjYWNoaW5nX3NoYTJfcGFzc3dvcmQifSx7Im5hbWUiOiJNWVNR"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzY5ODVjZjFhOWMwMjcyOTRlYmVmMjRkMzgyYjc4Yjk5OGFkZTk2M2QzNDQ3MDczMzk0OWJkYjRhMTdkZWI4NiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiODlkYjg1NzY4YTI2NmU3OWMyYTIyMWUyNjZlYzI4ZmVkN2FkNjI4MWI4Nzk4ZWJlNjJlYjhkOWRmZDRlOTAyOSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiODlkYjg1NzY4YTI2NmU3OWMyYTIyMWUyNjZlYzI4ZmVkN2FkNjI4MWI4Nzk4ZWJlNjJlYjhkOWRmZDRlOTAyOSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJwZXJjb25hLmNvbS9jb25maWd1cmF0aW9uLWhhc2giOiJkNDFkOGNkOThmMDBiMjA0ZTk4MDA5OThlY2Y4NDI3ZSIsInBlcmNvbmEuY29tL3NzbC1oYXNoIjoiY2ZkNmE1MjM5ODI2ODE3M2I1MWQ3Y2RiMTMzMWMwOWEiLCJwZXJjb25hLmNv"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNWQwMTUyZDQzYWE4ZjYyNjk5NGU0NzM0ZDc4YWIyMzlhMGYyNTYxMDQwNmMxYWUwOGJlMTI1OGE0MTA0ZmJiZSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJwZXJjb25hLmNvbS9jb25maWd1cmF0aW9uLWhhc2giOiJkNDFkOGNkOThmMDBiMjA0ZTk4MDA5OThlY2Y4NDI3ZSIsInBlcmNvbmEuY29tL3NzbC1oYXNoIjoiY2ZkNmE1MjM5ODI2ODE3M2I1MWQ3Y2RiMTMzMWMw"..., +  PeriodSeconds: 0, -  PeriodSeconds: 10, +  PersistentVolumeClaimRetentionPolicy: nil, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  Phase: "", -  Phase: "Pending", +  PodManagementPolicy: "", -  PodManagementPolicy: "OrderedReady", +  Protocol: "", -  Protocol: "TCP", +  ReadyReplicas: 0, -  ReadyReplicas: 2, -  ReadyReplicas: 3, +  Replicas: 0, -  Replicas: 2, -  Replicas: &2, +  Replicas: &2, -  Replicas: 3, -  Replicas: &3, +  Replicas: &3, +  ResourceVersion: "", -  ResourceVersion: "1764239815546575019", -  ResourceVersion: "1764240010009039008", -  ResourceVersion: "1764240147047839019", -  ResourceVersion: "1764240187273855019", -  ResourceVersion: "1764240202250031019", -  ResourceVersion: "1764240352959439008", -  ResourceVersion: "1764240402635487019", -  ResourceVersion: "1764240447704735019", -  ResourceVersion: "1764240503790543019", -  ResourceVersion: "1764240618330287008", -  ResourceVersion: "1764240667078335019", -  ResourceVersion: "1764240908836735008", -  ResourceVersion: "1764240918447455008", -  ResourceVersion: "1764240918496607003", -  ResourceVersion: "1764241017354127003", +  RestartPolicy: "", -  RestartPolicy: "Always", -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil, +  SchedulerName: "", +  SchedulerName: "", -  SchedulerName: "default-scheduler", -  SchedulerName: "default-scheduler", +  SecurityContext: nil, -  SecurityContext: s"&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,AppArmorProfile:nil,SupplementalGroupsPolicy:nil,SELinux"..., -  Subresource: "status", +  TerminationMessagePath: "", -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePolicy: "", -  TerminationMessagePolicy: "File", -  Time: s"2025-11-27 10:36:25 +0000 UTC", -  Time: s"2025-11-27 10:36:55 +0000 UTC", -  Time: s"2025-11-27 10:40:09 +0000 UTC", -  Time: s"2025-11-27 10:42:12 +0000 UTC", -  Time: s"2025-11-27 10:42:27 +0000 UTC", -  Time: s"2025-11-27 10:42:38 +0000 UTC", -  Time: s"2025-11-27 10:43:07 +0000 UTC", -  Time: s"2025-11-27 10:43:21 +0000 UTC", -  Time: s"2025-11-27 10:43:22 +0000 UTC", -  Time: s"2025-11-27 10:43:23 +0000 UTC", -  Time: s"2025-11-27 10:45:52 +0000 UTC", -  Time: s"2025-11-27 10:46:06 +0000 UTC", -  Time: s"2025-11-27 10:46:42 +0000 UTC", -  Time: s"2025-11-27 10:47:11 +0000 UTC", -  Time: s"2025-11-27 10:47:27 +0000 UTC", -  Time: s"2025-11-27 10:48:00 +0000 UTC", -  Time: s"2025-11-27 10:48:23 +0000 UTC", -  Time: s"2025-11-27 10:50:18 +0000 UTC", -  Time: s"2025-11-27 10:50:27 +0000 UTC", -  Time: s"2025-11-27 10:51:07 +0000 UTC", -  Time: s"2025-11-27 10:52:43 +0000 UTC", -  Time: s"2025-11-27 10:55:08 +0000 UTC", -  Time: s"2025-11-27 10:55:18 +0000 UTC", -  Time: s"2025-11-27 10:56:57 +0000 UTC", -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, +  UID: "", -  UID: "435ecf59-5eb4-4ae0-a705-fe351286dafe", -  UID: "b47dfa8b-f277-4fec-b6f9-0b6faa600ec3", -  UID: "dee202bf-cbb0-4238-9c4b-a76d2588e28f", +  UpdatedReplicas: 0, -  UpdatedReplicas: 2, -  UpdatedReplicas: 3, +  UpdateRevision: "", -  UpdateRevision: "some-name-haproxy-5b567bf96c", -  UpdateRevision: "some-name-proxysql-566898c866", -  UpdateRevision: "some-name-proxysql-5868b47f6b", -  UpdateRevision: "some-name-proxysql-65f5ccc684", -  UpdateRevision: "some-name-proxysql-746b976c8d", -  UpdateRevision: "some-name-proxysql-7965c48b57", -  UpdateRevision: "some-name-proxysql-f874d7649", -  UpdateRevision: "some-name-pxc-6b45b7d479", -  UpdateRevision: "some-name-pxc-766887bc84", -  UpdateRevision: "some-name-pxc-77d4f8d758", -  UpdateRevision: "some-name-pxc-7d5d5c48bb", -  VolumeMode: &"Filesystem", +  VolumeMode: nil, -  VolumeMounts: []v1.VolumeMount{{Name: "datadir", MountPath: "/var/lib/mysql"}},   }    },    },    {    },    },    {    },    }, ""),    },    {    },    },    },    ... // 16 identical fields    ... // 16 identical fields    "2",    ... // 22 identical fields    ... // 2 identical fields    ... // 2 identical fields    ... // 2 identical fields    ... // 2 identical fields    ... // 3 identical elements    ... // 3 identical fields    ... // 3 identical fields    ... // 3 identical fields    ... // 4 identical fields    "5",    ... // 5 identical elements    ... // 5 identical fields    ... // 5 identical fields    ... // 5 identical fields    "6",    ... // 6 identical fields    ... // 6 identical fields    ... // 7 identical fields    ... // 8 identical fields    ... // 9 identical fields    ... // 9 identical fields    AccessModes: nil,    ActiveDeadlineSeconds: nil,    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "haproxy", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}},    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}},    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}},    Annotations: map[string]string{    Args: {"haproxy"},    Args: {"mysqld"},    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    AutomountServiceAccountToken: nil,    AWSElasticBlockStore: nil,    AzureFile: nil,    Capacity: nil,    Conditions: nil,    ConfigMapKeyRef: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    ContainerPort: 3306,    ContainerPort: 33060,    ContainerPort: 33062,    ContainerPort: 3307,    ContainerPort: 3309,    ContainerPort: 4444,    ContainerPort: 4567,    ContainerPort: 4568,    ContainerPort: 6032,    ContainerPort: 6070,    ContainerPort: 8404,    Containers: []v1.Container{    DataSource: nil,    DataSourceRef: nil,    DeletionGracePeriodSeconds: nil,    DeletionGracePeriodSeconds: nil,    DeletionTimestamp: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-env-vars-haproxy"}, Optional: &true}}},    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-env-vars-proxysql"}, Optional: &true}}},    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-env-vars-pxc"}, Optional: &true}}},    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-log-collector"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "some-name-pxc"}, {Name: "LIVENESS_CHECK_TIMEOUT", Value: "5"}, {Name: "READINESS_CHECK_TIMEOUT", Value: "1"}},    Env: {{Name: "PXC_SERVICE", Value: "some-name-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "proxyadmin"}}}, ...},    Env: {{Name: "PXC_SERVICE", Value: "some-name-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "xtrabackup"}}}, ...},    Env: []v1.EnvVar{    EphemeralContainers: nil,    FailureThreshold: 3,    FC: nil,    FieldPath: "metadata.name",    FieldPath: "metadata.namespace",    FieldRef: &v1.ObjectFieldSelector{    Finalizers: nil,    Finalizers: nil,    GitRepo: nil,    HostAliases: nil,    HostAliases: nil,    HostIP: "",    HostIPC: false,    Hostname: "",    HostPort: 0,    ImagePullPolicy: "Always",    ImagePullSecrets: nil,    InitContainers: []v1.Container{    InitialDelaySeconds: 300,    ISCSI: nil,    Items: nil,    Items: nil,    "kubectl.kubernetes.io/default-container": "haproxy",    "kubectl.kubernetes.io/default-container": "proxysql",    "kubectl.kubernetes.io/default-container": "pxc",    Labels: {"app.kubernetes.io/component": "haproxy", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Labels: nil,    "last-applied-secret": strings.Join({    Lifecycle: nil,    LivenessProbe: &v1.Probe{    LocalObjectReference: {Name: "auto-some-name-pxc"},    LocalObjectReference: {Name: "some-name-haproxy"},    LocalObjectReference: {Name: "some-name-pxc"},    ManagedFields: nil,    MinReadySeconds: 0,    Name: "auto-config",    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    {Name: "CLUSTER_HASH", Value: "1867224"},    Name: "config",    {Name: "haproxy-auto", VolumeSource: {EmptyDir: &{}}},    Name: "haproxy-custom",    Name: "ist",    {Name: "LOG_DATA_DIR", Value: "/var/lib/mysql"},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "monitor"}}},    Name: "mysql",    Name: "mysql-admin",    Name: "mysql-init-file",    Name: "mysql-replicas",    Name: "mysql-users-secret-file",    Name: "mysqlx",    {Name: "OPERATOR_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "operator"}}},    Name: "POD_NAME",    Name: "POD_NAMESPASE",    Name: "proxyadm",    Name: "proxy-protocol",    Name: "some-name-env-vars-haproxy",    Namespace: "users-3672",    Name: "ssl",    Name: "ssl-internal",    Name: "sst",    Name: "stats",    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    Name: "vault-keyring-secret",    Name: "write-set",    {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "xtrabackup"}}},    NFS: nil,    NodeName: "",    NodeSelector: nil,    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "last-applied-secret": "5d0152d43aa8f626994e4734d78ab239a0f25610406c1ae08be1258a4104fbbe", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", ...}},    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc"}},    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "352ec183c1b6b8ef93cea415a5bd6f047e75886a57fdd8d864e7cd06f4528aed", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", ...}},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    ObjectMeta: v1.ObjectMeta{    ObjectMeta: v1.ObjectMeta{    Optional: &false,    Optional: &true,    Optional: &true,    Ordinals: nil,    OS: nil,    Overhead: nil,    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "some-name", UID: "901448f5-b79e-4d95-99ad-6a96d88705df", ...}},    OwnerReferences: nil,    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a",    "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a",    "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc",    "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc",    Ports: nil,    Ports: []v1.ContainerPort{    PreemptionPolicy: nil,    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    Quobyte: nil,    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    Replicas: &2,    Replicas: &3,    ResizePolicy: nil,    ResourceFieldRef: nil,    Resources: {},    Resources: {Limits: {s"cpu": {i: {...}, s: "1", Format: "DecimalSI"}, s"memory": {i: {...}, s: "2G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "DecimalSI"}}},    SecretName: "internal-some-name",    SecretName: "some-name-env-vars-haproxy",    SecretName: "some-name-mysql-init",    SecretName: "some-name-ssl",    SecretName: "some-name-ssl-internal",    SecretName: "some-name-vault",    Secret: &v1.SecretVolumeSource{    SecurityContext: nil,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "haproxy", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    SelfLink: "",    ServiceAccountName: "default",    ServiceName: "some-name-haproxy",    ServiceName: "some-name-proxysql-unready",    ServiceName: "some-name-pxc",    SetHostnameAsFQDN: nil,    ShareProcessNamespace: nil,    Spec: v1.PersistentVolumeClaimSpec{    Spec: v1.PodSpec{    Spec: v1.StatefulSetSpec{    StartupProbe: nil,    Status: {},    Status: v1.PersistentVolumeClaimStatus{    Status: v1.StatefulSetStatus{    StorageClassName: nil,    Subdomain: "",    Subdomain: "",    SuccessThreshold: 1,    Template: v1.PodTemplateSpec{    TerminationGracePeriodSeconds: &30,    TerminationGracePeriodSeconds: &600,    TerminationGracePeriodSeconds: nil,    TimeoutSeconds: 5,    Tolerations: {{Key: "node.alpha.kubernetes.io/unreachable", Operator: "Exists", Effect: "NoExecute", TolerationSeconds: &6000}},    Tolerations: nil,    TypeMeta: {},    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}},   &v1.StatefulSet{    Value: "",    ValueFrom: &v1.EnvVarSource{    VolumeAttributesClassName: nil,    VolumeClaimTemplates: nil,    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    VolumeDevices: nil,    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeName: "",    VolumeSource: v1.VolumeSource{    Volumes: []v1.Volume{    VsphereVolume: nil,    WorkingDir: "", + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n users-3672 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.VQPhqXZ9E6 ++ mktemp + local LAST_ERR=/tmp/tmp.bUJHYb3in2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VQPhqXZ9E6 perconaxtradbcluster.pxc.percona.com "some-name" deleted from users-3672 namespace + cat /tmp/tmp.bUJHYb3in2 + rm /tmp/tmp.VQPhqXZ9E6 /tmp/tmp.bUJHYb3in2 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.LWrtGoZiZV ++ mktemp + local LAST_ERR=/tmp/tmp.kPAepWoF8S + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LWrtGoZiZV No resources found + cat /tmp/tmp.kPAepWoF8S + rm /tmp/tmp.LWrtGoZiZV /tmp/tmp.kPAepWoF8S + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.ytPo80jqu6 ++ mktemp + local LAST_ERR=/tmp/tmp.ZIT8mHBWUF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ytPo80jqu6 No resources found + cat /tmp/tmp.ZIT8mHBWUF + rm /tmp/tmp.ytPo80jqu6 /tmp/tmp.ZIT8mHBWUF + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.ddugqzEkOd ++ mktemp + local LAST_ERR=/tmp/tmp.La54Pj7K5U + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ddugqzEkOd validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.La54Pj7K5U + rm /tmp/tmp.ddugqzEkOd /tmp/tmp.La54Pj7K5U + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace users-3672 + rm -rf /tmp/tmp.miV4rrH5zD + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.CcV7UMDabA + desc 'test passed' + set +o xtrace ++ mktemp ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.amGdWu3CA8 + local LAST_ERR=/tmp/tmp.4D9aZQf5iw + local exit_status=0 ++ mktemp + local LAST_ERR=/tmp/tmp.YWHFqCcp8b + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace users-3672 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator