Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/logs/users-scheduler-8-4.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + create_infra users-scheduler-5913 + local ns=users-scheduler-5913 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified + : + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.SpJiGB7VbA ++ mktemp + local LAST_ERR=/tmp/tmp.5oisIkm5G9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SpJiGB7VbA No resources found + cat /tmp/tmp.5oisIkm5G9 + rm /tmp/tmp.SpJiGB7VbA /tmp/tmp.5oisIkm5G9 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.zLolN9pIN6 ++ mktemp + local LAST_ERR=/tmp/tmp.cgNT29bpps + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zLolN9pIN6 No resources found + cat /tmp/tmp.cgNT29bpps + rm /tmp/tmp.zLolN9pIN6 /tmp/tmp.cgNT29bpps + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.eRSQh16DV6 ++ mktemp + local LAST_ERR=/tmp/tmp.sobGSGHckH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eRSQh16DV6 No resources found + cat /tmp/tmp.sobGSGHckH + rm /tmp/tmp.eRSQh16DV6 /tmp/tmp.sobGSGHckH + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// ++ tail -n1 ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + xargs kubectl delete ns + kubectl_bin get ns + local LAST_OUT=/tmp/tmp.4QtItWUpju ++ mktemp + local LAST_ERR=/tmp/tmp.ZSNjezLLG9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.vTV94XEV6c ++ mktemp + local LAST_ERR=/tmp/tmp.JGbWKLPfGJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vTV94XEV6c + cat /tmp/tmp.JGbWKLPfGJ + rm /tmp/tmp.vTV94XEV6c /tmp/tmp.JGbWKLPfGJ + return 0 namespace "users-scheduler-2774" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4QtItWUpju namespace "pxc-operator" deleted + cat /tmp/tmp.ZSNjezLLG9 + rm /tmp/tmp.4QtItWUpju /tmp/tmp.ZSNjezLLG9 + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.rz4HA0p8Nq ++ mktemp + localerror: timed out waiting for the condition on pods/some-name-proxysql-0 waiting for pod/some-name-proxysql-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ERROR 1062 (23000) at line 1: Duplicate entry '100500' for key 'myApp.PRIMARY' command terminated with exit code 1 pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok --- /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1.sql 2026-04-09 01:07:44.955099464 +0000 +++ /tmp/tmp.uAchwulIlq/select-1.sql 2026-04-09 05:27:49.436663447 +0000 @@ -1 +1,2 @@ -100500 +ERROR 2003 (HY000): Can't connect to MySQL server on 'some-name-pxc-2.some-name-pxc:3306' (111) +command terminated with exit code 1 break + cat /tmp/tmp.dZoCLziRKO customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.GmhxhBOS4j + rm /tmp/tmp.dZoCLziRKO /tmp/tmp.GmhxhBOS4j + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + sed -e 's^namespace: .*^namespace: pxc-operator^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/deploy/cw-rbac.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.9z0CDZHEPU ++ mktemp + local LAST_ERR=/tmp/tmp.jsguMSz6hs + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9z0CDZHEPU clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.jsguMSz6hs + rm /tmp/tmp.9z0CDZHEPU /tmp/tmp.jsguMSz6hs + return 0 + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2434-3b65d7fe^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = ""' - + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.HV3NQDqXh0 ++ mktemp + local LAST_ERR=/tmp/tmp.GSzvXW9JIS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HV3NQDqXh0 deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.GSzvXW9JIS + rm /tmp/tmp.HV3NQDqXh0 /tmp/tmp.GSzvXW9JIS + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.qhbhwontiQ ++ mktemp + local LAST_ERR=/tmp/tmp.xbnezzcPf8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qhbhwontiQ pod/percona-xtradb-cluster-operator-665ff7485-p85m7 condition met + cat /tmp/tmp.xbnezzcPf8 + rm /tmp/tmp.qhbhwontiQ /tmp/tmp.xbnezzcPf8 + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ head -1 +++ mktemp ++ local LAST_OUT=/tmp/tmp.ntjvJKrX2x +++ mktemp ++ local LAST_ERR=/tmp/tmp.cudTZZApAk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ntjvJKrX2x ++ cat /tmp/tmp.cudTZZApAk ++ rm /tmp/tmp.ntjvJKrX2x /tmp/tmp.cudTZZApAk ++ return 0 + wait_pod percona-xtradb-cluster-operator-665ff7485-p85m7 480 pxc-operator + local pod=percona-xtradb-cluster-operator-665ff7485-p85m7 + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-665ff7485-p85m7 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-665ff7485-p85m7 condition met waiting for pod/percona-xtradb-cluster-operator-665ff7485-p85m7 to become Ready.Ok + sleep 3 + create_namespace users-scheduler-5913 + local namespace=users-scheduler-5913 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces users-scheduler-5913' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces users-scheduler-5913 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace users-scheduler-5913 + kubectl_bin get ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.PF1DePtRuS + local LAST_OUT=/tmp/tmp.ToVLNCbgox ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.n3aSvF4JxJ + local exit_status=0 + local LAST_ERR=/tmp/tmp.oG6WSlQmxC + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + for i in '$(seq 0 2)' + set +e + kubectl get ns + kubectl delete namespace users-scheduler-5913 + xargs kubectl delete ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-scheduler-5913 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ToVLNCbgox + cat /tmp/tmp.n3aSvF4JxJ + rm /tmp/tmp.ToVLNCbgox /tmp/tmp.n3aSvF4JxJ + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace users-scheduler-5913 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.PF1DePtRuS + cat /tmp/tmp.oG6WSlQmxC Error from server (NotFound): namespaces "users-scheduler-5913" not found + rm /tmp/tmp.PF1DePtRuS /tmp/tmp.oG6WSlQmxC + return 1 + : + wait_for_delete namespace/users-scheduler-5913 + local res=namespace/users-scheduler-5913 + echo -n 'waiting for namespace/users-scheduler-5913 to be deleted' waiting for namespace/users-scheduler-5913 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "users-scheduler-5913" not found + desc 'create namespace users-scheduler-5913' + set +o xtrace ----------------------------------------------------------------------------------- create namespace users-scheduler-5913 ----------------------------------------------------------------------------------- + kubectl_bin create namespace users-scheduler-5913 ++ mktemp + local LAST_OUT=/tmp/tmp.ONnjfWDkDO ++ mktemp + local LAST_ERR=/tmp/tmp.o2wH03uh1T + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace users-scheduler-5913 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ONnjfWDkDO namespace/users-scheduler-5913 created + cat /tmp/tmp.o2wH03uh1T + rm /tmp/tmp.ONnjfWDkDO /tmp/tmp.o2wH03uh1T + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.AK7TwAMFSM +++ mktemp ++ local LAST_ERR=/tmp/tmp.mGDwM4czBu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AK7TwAMFSM ++ cat /tmp/tmp.mGDwM4czBu ++ rm /tmp/tmp.AK7TwAMFSM /tmp/tmp.mGDwM4czBu ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2434-3b65d7fe-1-cluster6 --namespace=users-scheduler-5913 ++ mktemp + local LAST_OUT=/tmp/tmp.zMmtam9jou ++ mktemp + local LAST_ERR=/tmp/tmp.nECsnA8HuQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2434-3b65d7fe-1-cluster6 --namespace=users-scheduler-5913 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zMmtam9jou Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2434-3b65d7fe-1-cluster6" modified. + cat /tmp/tmp.nECsnA8HuQ + rm /tmp/tmp.zMmtam9jou /tmp/tmp.nECsnA8HuQ + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9gWF57Kh8x ++ mktemp + local LAST_ERR=/tmp/tmp.jdaVhloWMF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9gWF57Kh8x secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.jdaVhloWMF + rm /tmp/tmp.9gWF57Kh8x /tmp/tmp.jdaVhloWMF + return 0 + desc 'create PXC cluster with 1-password secret' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster with 1-password secret ----------------------------------------------------------------------------------- + newpass=test-password ++ echo -n test-password ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZA== + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/some-name.yml '' '' /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/secrets_one_pass.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/secrets_one_pass.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/secrets_one_pass.yml ++ mktemp + local LAST_OUT=/tmp/tmp.QSGrEiJlCa ++ mktemp + local LAST_ERR=/tmp/tmp.AqGw6tCGuf + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/secrets_one_pass.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QSGrEiJlCa secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.AqGw6tCGuf + rm /tmp/tmp.QSGrEiJlCa /tmp/tmp.AqGw6tCGuf + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/client.yml + local pvc_name= + kubectl_bin apply -f - + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' ++ mktemp + local LAST_OUT=/tmp/tmp.w0ZwYegKmi + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-scheduler-5913~ + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/conf/client.yml ++ mktemp + local LAST_ERR=/tmp/tmp.rN9VkKsBIz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2434-3b65d7fe#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.w0ZwYegKmi deployment.apps/pxc-client created + cat /tmp/tmp.rN9VkKsBIz + rm /tmp/tmp.w0ZwYegKmi /tmp/tmp.rN9VkKsBIz + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/some-name.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/some-name.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/some-name.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/some-name.yml + local pvc_name= + kubectl_bin apply -f - + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_OUT=/tmp/tmp.RUvGIuNkyN + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-scheduler-5913~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2434-3b65d7fe#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' ++ mktemp + local LAST_ERR=/tmp/tmp.xKBN0JQNmM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RUvGIuNkyN perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.xKBN0JQNmM + rm /tmp/tmp.RUvGIuNkyN /tmp/tmp.xKBN0JQNmM + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.R8dV9zNttI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Q1DQtSq2Ft +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.R8dV9zNttI +++ cat /tmp/tmp.Q1DQtSq2Ft +++ rm /tmp/tmp.R8dV9zNttI /tmp/tmp.Q1DQtSq2Ft +++ return 0 ++ [[ false == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Qm6RxiL3Yz ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5dzVqJyNpt +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Qm6RxiL3Yz +++ cat /tmp/tmp.5dzVqJyNpt +++ rm /tmp/tmp.Qm6RxiL3Yz /tmp/tmp.5dzVqJyNpt +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-5913 ++ mktemp + local LAST_OUT=/tmp/tmp.xFpisoml7R ++ mktemp + local LAST_ERR=/tmp/tmp.cJTqULGbtT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-5913 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-5913 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-5913 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.xFpisoml7R + cat /tmp/tmp.cJTqULGbtT error: no matching resources found + rm /tmp/tmp.xFpisoml7R /tmp/tmp.cJTqULGbtT + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-0 + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.3jfGBrgOL0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qsAkuHur1V ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3jfGBrgOL0 ++ cat /tmp/tmp.qsAkuHur1V ++ rm /tmp/tmp.3jfGBrgOL0 /tmp/tmp.qsAkuHur1V ++ return 0 + local 'root_pass=q-EeL6hn~{EYk<[A' + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Du1tLMk0RQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.tiPeC3ARLf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Du1tLMk0RQ ++ cat /tmp/tmp.tiPeC3ARLf ++ rm /tmp/tmp.Du1tLMk0RQ /tmp/tmp.tiPeC3ARLf ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qcQ49WDpe7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yTxDjuh8JH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qcQ49WDpe7 ++ cat /tmp/tmp.yTxDjuh8JH ++ rm /tmp/tmp.qcQ49WDpe7 /tmp/tmp.yTxDjuh8JH ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AWDmABsjrD +++ mktemp ++ local LAST_ERR=/tmp/tmp.yMx5qbXob0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AWDmABsjrD ++ cat /tmp/tmp.yMx5qbXob0 ++ rm /tmp/tmp.AWDmABsjrD /tmp/tmp.yMx5qbXob0 ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1.sql /tmp/tmp.QDkOK88Jf9/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V6ZgFtokgv +++ mktemp ++ local LAST_ERR=/tmp/tmp.FTuegB3VtM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.V6ZgFtokgv ++ cat /tmp/tmp.FTuegB3VtM ++ rm /tmp/tmp.V6ZgFtokgv /tmp/tmp.FTuegB3VtM ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1.sql /tmp/tmp.QDkOK88Jf9/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''q-EeL6hn~{EYk<[A'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vCCTEUcybV +++ mktemp ++ local LAST_ERR=/tmp/tmp.sVau8qhfpJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vCCTEUcybV ++ cat /tmp/tmp.sVau8qhfpJ ++ rm /tmp/tmp.vCCTEUcybV /tmp/tmp.sVau8qhfpJ ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-1.sql /tmp/tmp.QDkOK88Jf9/select-1.sql + is_keyring_plugin_in_use some-name + local cluster=some-name + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + kubectl exec some-name-pxc-0 -c pxc -- ls /var/lib/mysql/mysqld.my + return 1 ++ kubectl exec -it some-name-proxysql-0 -- sh -c 'proxysql --version 2>/dev/null' ++ awk '{print $3}' ++ cut -d. -f1 Unable to use a TTY - input is not a terminal or the right kind of file + PROXYSQL_VER=2 + tables_cmp_file=select-2-proxysql3 + [[ 2 == 2 ]] + tables_cmp_file=select-2 + desc 'test missing passwords were created and present in internal secrets' + set +o xtrace ----------------------------------------------------------------------------------- test missing passwords were created and present in internal secrets ----------------------------------------------------------------------------------- + empty_pwds=() + wrong_pwds=() + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking root' Checking root ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.8ZwwadtKGj +++ mktemp ++ local LAST_ERR=/tmp/tmp.PBZEuMeH57 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8ZwwadtKGj ++ cat /tmp/tmp.PBZEuMeH57 ++ rm /tmp/tmp.8ZwwadtKGj /tmp/tmp.PBZEuMeH57 ++ return 0 + secret_pass='q-EeL6hn~{EYk<[A' ++ getSecretData internal-some-name root ++ local secretName=internal-some-name ++ local dataKey=root ++ kubectl_bin get secrets/internal-some-name '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k7YfVS0uKq +++ mktemp ++ local LAST_ERR=/tmp/tmp.XZwfXcJcBA ++ local exit_status=0 +++ seq 0 2 ++ base64 --decode ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k7YfVS0uKq ++ cat /tmp/tmp.XZwfXcJcBA ++ rm /tmp/tmp.k7YfVS0uKq /tmp/tmp.XZwfXcJcBA ++ return 0 + int_secret_pass='q-EeL6hn~{EYk<[A' + [[ -z q-EeL6hn~{EYk<[A ]] + [[ q-EeL6hn~{EYk<[A != \q\-\E\e\L\6\h\n\~\{\E\Y\k\<\[\A ]] + [[ root != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ root ]] + [[ '' =~ root ]] + echo 'Running compare for root' Running compare for root + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''q-EeL6hn~{EYk<[A'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JrGKGODmmO +++ mktemp ++ local LAST_ERR=/tmp/tmp.UYeS64qrsi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JrGKGODmmO ++ cat /tmp/tmp.UYeS64qrsi ++ rm /tmp/tmp.JrGKGODmmO /tmp/tmp.UYeS64qrsi ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking xtrabackup' Checking xtrabackup ++ getSecretData my-cluster-secrets xtrabackup ++ local secretName=my-cluster-secrets ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.jZ4wrKHt6C +++ mktemp ++ local LAST_ERR=/tmp/tmp.ylCClEae1A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jZ4wrKHt6C ++ cat /tmp/tmp.ylCClEae1A ++ rm /tmp/tmp.jZ4wrKHt6C /tmp/tmp.ylCClEae1A ++ return 0 + secret_pass='0O(i=k*F$uLe)dSYr' ++ getSecretData internal-some-name xtrabackup ++ local secretName=internal-some-name ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.DfHx2YvH5P +++ mktemp ++ local LAST_ERR=/tmp/tmp.M8D8dPgNm4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DfHx2YvH5P ++ cat /tmp/tmp.M8D8dPgNm4 ++ rm /tmp/tmp.DfHx2YvH5P /tmp/tmp.M8D8dPgNm4 ++ return 0 + int_secret_pass='0O(i=k*F$uLe)dSYr' + [[ -z 0O(i=k*F$uLe)dSYr ]] + [[ 0O(i=k*F$uLe)dSYr != \0\O\(\i\=\k\*\F\$\u\L\e\)\d\S\Y\r ]] + [[ xtrabackup != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ xtrabackup ]] + [[ '' =~ xtrabackup ]] + echo 'Running compare for xtrabackup' Running compare for xtrabackup + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\''0O(i=k*F$uLe)dSYr'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\''0O(i=k*F$uLe)dSYr'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\''0O(i=k*F$uLe)dSYr'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\''0O(i=k*F$uLe)dSYr'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9kM3vVqHuO +++ mktemp ++ local LAST_ERR=/tmp/tmp.1T6n0fJdct ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9kM3vVqHuO ++ cat /tmp/tmp.1T6n0fJdct ++ rm /tmp/tmp.9kM3vVqHuO /tmp/tmp.1T6n0fJdct ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking monitor' Checking monitor ++ getSecretData my-cluster-secrets monitor ++ local secretName=my-cluster-secrets ++ local dataKey=monitor ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.AswxRgxD8J +++ mktemp ++ local LAST_ERR=/tmp/tmp.L0O1jFDDwY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AswxRgxD8J ++ cat /tmp/tmp.L0O1jFDDwY ++ rm /tmp/tmp.AswxRgxD8J /tmp/tmp.L0O1jFDDwY ++ return 0 + secret_pass=monitor_password ++ getSecretData internal-some-name monitor ++ local secretName=internal-some-name ++ local dataKey=monitor ++ kubectl_bin get secrets/internal-some-name '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.nVO5fdEqwQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.wgTmaDd8Vi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nVO5fdEqwQ ++ cat /tmp/tmp.wgTmaDd8Vi ++ rm /tmp/tmp.nVO5fdEqwQ /tmp/tmp.wgTmaDd8Vi ++ return 0 + int_secret_pass=monitor_password + [[ -z monitor_password ]] + [[ monitor_password != \m\o\n\i\t\o\r\_\p\a\s\s\w\o\r\d ]] + [[ monitor != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ monitor ]] + [[ '' =~ monitor ]] + echo 'Running compare for monitor' Running compare for monitor + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BUU94XLquO +++ mktemp ++ local LAST_ERR=/tmp/tmp.GpkIdrdBTk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BUU94XLquO ++ cat /tmp/tmp.GpkIdrdBTk ++ rm /tmp/tmp.BUU94XLquO /tmp/tmp.GpkIdrdBTk ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking proxyadmin' Checking proxyadmin ++ getSecretData my-cluster-secrets proxyadmin ++ local secretName=my-cluster-secrets ++ local dataKey=proxyadmin ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.8iLz1o9Vek +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZJdV7mdbSZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8iLz1o9Vek ++ cat /tmp/tmp.ZJdV7mdbSZ ++ rm /tmp/tmp.8iLz1o9Vek /tmp/tmp.ZJdV7mdbSZ ++ return 0 + secret_pass='2]6}EE4mr>x^Fw^B' ++ getSecretData internal-some-name proxyadmin ++ local secretName=internal-some-name ++ local dataKey=proxyadmin ++ kubectl_bin get secrets/internal-some-name '--template={{.data.proxyadmin}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.uOn2KtfOXV +++ mktemp ++ local LAST_ERR=/tmp/tmp.R68eHaGCjm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.proxyadmin}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uOn2KtfOXV ++ cat /tmp/tmp.R68eHaGCjm ++ rm /tmp/tmp.uOn2KtfOXV /tmp/tmp.R68eHaGCjm ++ return 0 + int_secret_pass='2]6}EE4mr>x^Fw^B' + [[ -z 2]6}EE4mr>x^Fw^B ]] + [[ 2]6}EE4mr>x^Fw^B != \2\]\6\}\E\E\4\m\r\>\x\^\F\w\^\B ]] + [[ proxyadmin != \p\r\o\x\y\a\d\m\i\n ]] + [[ proxyadmin == \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ proxyadmin ]] + [[ '' =~ proxyadmin ]] + echo 'Running compare for proxyadmin' Running compare for proxyadmin + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''2]6}EE4mr>x^Fw^B'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''2]6}EE4mr>x^Fw^B'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''2]6}EE4mr>x^Fw^B'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''2]6}EE4mr>x^Fw^B'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.QDkOK88Jf9/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.QDkOK88Jf9/select-2.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking operator' Checking operator ++ getSecretData my-cluster-secrets operator ++ local secretName=my-cluster-secrets ++ local dataKey=operator ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.oKoywvF3gR +++ mktemp ++ local LAST_ERR=/tmp/tmp.Tzix36o81A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oKoywvF3gR ++ cat /tmp/tmp.Tzix36o81A ++ rm /tmp/tmp.oKoywvF3gR /tmp/tmp.Tzix36o81A ++ return 0 + secret_pass=_bBCbArFfAp5TR0,mDI ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' +++ mktemp ++ base64 --decode ++ local LAST_OUT=/tmp/tmp.tTJLcL0jiM +++ mktemp ++ local LAST_ERR=/tmp/tmp.RW1ChvW9Gg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tTJLcL0jiM ++ cat /tmp/tmp.RW1ChvW9Gg ++ rm /tmp/tmp.tTJLcL0jiM /tmp/tmp.RW1ChvW9Gg ++ return 0 + int_secret_pass=_bBCbArFfAp5TR0,mDI + [[ -z _bBCbArFfAp5TR0,mDI ]] + [[ _bBCbArFfAp5TR0,mDI != \_\b\B\C\b\A\r\F\f\A\p\5\T\R\0\,\m\D\I ]] + [[ operator != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ operator ]] + [[ '' =~ operator ]] + echo 'Running compare for operator' Running compare for operator + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''_bBCbArFfAp5TR0,mDI'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''_bBCbArFfAp5TR0,mDI'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''_bBCbArFfAp5TR0,mDI'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''_bBCbArFfAp5TR0,mDI'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2o6nU044uX +++ mktemp ++ local LAST_ERR=/tmp/tmp.BpJ16X0hxW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2o6nU044uX ++ cat /tmp/tmp.BpJ16X0hxW ++ rm /tmp/tmp.2o6nU044uX /tmp/tmp.BpJ16X0hxW ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking replication' Checking replication ++ getSecretData my-cluster-secrets replication ++ local secretName=my-cluster-secrets ++ local dataKey=replication ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.replication}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.RLo4i1AEct +++ mktemp ++ local LAST_ERR=/tmp/tmp.UAHDGaOvzw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.replication}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RLo4i1AEct ++ cat /tmp/tmp.UAHDGaOvzw ++ rm /tmp/tmp.RLo4i1AEct /tmp/tmp.UAHDGaOvzw ++ return 0 + secret_pass=')*QQ8u0+Fj~={3{Rw' ++ getSecretData internal-some-name replication ++ local secretName=internal-some-name ++ local dataKey=replication ++ kubectl_bin get secrets/internal-some-name '--template={{.data.replication}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.CMrc0ntT5V +++ mktemp ++ local LAST_ERR=/tmp/tmp.DxdzU3BVfI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.replication}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CMrc0ntT5V ++ cat /tmp/tmp.DxdzU3BVfI ++ rm /tmp/tmp.CMrc0ntT5V /tmp/tmp.DxdzU3BVfI ++ return 0 + int_secret_pass=')*QQ8u0+Fj~={3{Rw' + [[ -z )*QQ8u0+Fj~={3{Rw ]] + [[ )*QQ8u0+Fj~={3{Rw != \)\*\Q\Q\8\u\0\+\F\j\~\=\{\3\{\R\w ]] + [[ replication != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ replication ]] + [[ '' =~ replication ]] + echo 'Running compare for replication' Running compare for replication + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\'')*QQ8u0+Fj~={3{Rw'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\'')*QQ8u0+Fj~={3{Rw'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\'')*QQ8u0+Fj~={3{Rw'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\'')*QQ8u0+Fj~={3{Rw'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TO2KitvxSQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.esRog6iAzR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TO2KitvxSQ ++ cat /tmp/tmp.esRog6iAzR ++ rm /tmp/tmp.TO2KitvxSQ /tmp/tmp.esRog6iAzR ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ -n '' ]] + desc 'test root' + set +o xtrace ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets root dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=root + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.dYwgRPBHJO ++ mktemp + local LAST_ERR=/tmp/tmp.7hlkqvkxFR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dYwgRPBHJO secret/my-cluster-secrets patched + cat /tmp/tmp.7hlkqvkxFR + rm /tmp/tmp.dYwgRPBHJO /tmp/tmp.7hlkqvkxFR + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i8Q4iH1UTY +++ mktemp ++ local LAST_ERR=/tmp/tmp.hiXvFoEK77 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.i8Q4iH1UTY ++ cat /tmp/tmp.hiXvFoEK77 ++ rm /tmp/tmp.i8Q4iH1UTY /tmp/tmp.hiXvFoEK77 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F9xxHaRcgU +++ mktemp ++ local LAST_ERR=/tmp/tmp.w6iQV9b474 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.F9xxHaRcgU ++ cat /tmp/tmp.w6iQV9b474 ++ rm /tmp/tmp.F9xxHaRcgU /tmp/tmp.w6iQV9b474 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.19Hrr9Cwfm ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.fIs0fcp5gW +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.19Hrr9Cwfm +++++ cat /tmp/tmp.fIs0fcp5gW +++++ rm /tmp/tmp.19Hrr9Cwfm /tmp/tmp.fIs0fcp5gW +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.lby1z0lyxP ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.toHdmjDrak +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.lby1z0lyxP +++++ cat /tmp/tmp.toHdmjDrak +++++ rm /tmp/tmp.lby1z0lyxP /tmp/tmp.toHdmjDrak +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ebdzi7vLOM +++ mktemp ++ local LAST_ERR=/tmp/tmp.N6d1OxF0Rf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ebdzi7vLOM ++ cat /tmp/tmp.N6d1OxF0Rf ++ rm /tmp/tmp.Ebdzi7vLOM /tmp/tmp.N6d1OxF0Rf ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6U2fcY0Fn6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Cyhzn8UVrw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6U2fcY0Fn6 ++ cat /tmp/tmp.Cyhzn8UVrw ++ rm /tmp/tmp.6U2fcY0Fn6 /tmp/tmp.Cyhzn8UVrw ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + desc 'test proxyadmin' + set +o xtrace ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.i0YlkoMv5i ++ mktemp + local LAST_ERR=/tmp/tmp.dy25WMkiVx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.i0YlkoMv5i perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.dy25WMkiVx + rm /tmp/tmp.i0YlkoMv5i /tmp/tmp.dy25WMkiVx + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Gn31cCGmvd +++ mktemp ++ local LAST_ERR=/tmp/tmp.UgEJljJRzO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Gn31cCGmvd ++ cat /tmp/tmp.UgEJljJRzO ++ rm /tmp/tmp.Gn31cCGmvd /tmp/tmp.UgEJljJRzO ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.trjHMGDvxX +++ mktemp ++ local LAST_ERR=/tmp/tmp.SBNUHx5TUm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.trjHMGDvxX ++ cat /tmp/tmp.SBNUHx5TUm ++ rm /tmp/tmp.trjHMGDvxX /tmp/tmp.SBNUHx5TUm ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.3SJ1XnyEGJ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.dRgK2w3i21 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.3SJ1XnyEGJ +++++ cat /tmp/tmp.dRgK2w3i21 +++++ rm /tmp/tmp.3SJ1XnyEGJ /tmp/tmp.dRgK2w3i21 +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.eQj9zibQnW ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.2RlTgHmBii +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.eQj9zibQnW +++++ cat /tmp/tmp.2RlTgHmBii +++++ rm /tmp/tmp.eQj9zibQnW /tmp/tmp.2RlTgHmBii +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.24KesRaKxM +++ mktemp ++ local LAST_ERR=/tmp/tmp.4ew6a7Fph7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.24KesRaKxM ++ cat /tmp/tmp.4ew6a7Fph7 ++ rm /tmp/tmp.24KesRaKxM /tmp/tmp.4ew6a7Fph7 ++ return 0 + [[ 3 == \3 ]] + echo + patch_secret my-cluster-secrets proxyadmin dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=proxyadmin + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.tHDCa3x1Jm ++ mktemp + local LAST_ERR=/tmp/tmp.IQPdWyqW6W + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tHDCa3x1Jm secret/my-cluster-secrets patched + cat /tmp/tmp.IQPdWyqW6W + rm /tmp/tmp.tHDCa3x1Jm /tmp/tmp.IQPdWyqW6W + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GAhKTwLkz2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.iGw7kHbLEG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GAhKTwLkz2 ++ cat /tmp/tmp.iGw7kHbLEG ++ rm /tmp/tmp.GAhKTwLkz2 /tmp/tmp.iGw7kHbLEG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2alwvYKNTL +++ mktemp ++ local LAST_ERR=/tmp/tmp.bQh58dCr3V ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2alwvYKNTL ++ cat /tmp/tmp.bQh58dCr3V ++ rm /tmp/tmp.2alwvYKNTL /tmp/tmp.bQh58dCr3V ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PdA9uRwbih +++ mktemp ++ local LAST_ERR=/tmp/tmp.bXiYm2DakO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PdA9uRwbih ++ cat /tmp/tmp.bXiYm2DakO ++ rm /tmp/tmp.PdA9uRwbih /tmp/tmp.bXiYm2DakO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Uf6ONtbR6w +++ mktemp ++ local LAST_ERR=/tmp/tmp.dpV05IGSbJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Uf6ONtbR6w ++ cat /tmp/tmp.dpV05IGSbJ ++ rm /tmp/tmp.Uf6ONtbR6w /tmp/tmp.dpV05IGSbJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sQvQrpiBnP +++ mktemp ++ local LAST_ERR=/tmp/tmp.mpZ8j4inQT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sQvQrpiBnP ++ cat /tmp/tmp.mpZ8j4inQT ++ rm /tmp/tmp.sQvQrpiBnP /tmp/tmp.mpZ8j4inQT ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1LlxGxxOz4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.i5DJdOe5Dy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1LlxGxxOz4 ++ cat /tmp/tmp.i5DJdOe5Dy ++ rm /tmp/tmp.1LlxGxxOz4 /tmp/tmp.i5DJdOe5Dy ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.B0SZl0w2Cr ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.6Mw6ED2HSB +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.B0SZl0w2Cr +++++ cat /tmp/tmp.6Mw6ED2HSB +++++ rm /tmp/tmp.B0SZl0w2Cr /tmp/tmp.6Mw6ED2HSB +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.kkPtmlWvaJ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.05oNgqQbkm +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.kkPtmlWvaJ +++++ cat /tmp/tmp.05oNgqQbkm +++++ rm /tmp/tmp.kkPtmlWvaJ /tmp/tmp.05oNgqQbkm +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4TQihHlKJW +++ mktemp ++ local LAST_ERR=/tmp/tmp.PFYMeQNvpB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4TQihHlKJW ++ cat /tmp/tmp.PFYMeQNvpB ++ rm /tmp/tmp.4TQihHlKJW /tmp/tmp.PFYMeQNvpB ++ return 0 + [[ 3 == \3 ]] + echo + sleep 15 + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.QDkOK88Jf9/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.QDkOK88Jf9/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.QDkOK88Jf9/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.QDkOK88Jf9/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.QDkOK88Jf9/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.QDkOK88Jf9/select-2.sql + compare_mysql_cmd_local proxysql-cfg 'SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=proxysql-cfg + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/proxysql-cfg.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/proxysql-cfg-84.sql ']' + run_mysql_local 'SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.QDkOK88Jf9/proxysql-cfg.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/proxysql-cfg.sql /tmp/tmp.QDkOK88Jf9/proxysql-cfg.sql + desc 'test xtrabackup' + set +o xtrace ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.PzkvXvsGNo ++ mktemp + local LAST_ERR=/tmp/tmp.cP46vE7cCB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PzkvXvsGNo perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.cP46vE7cCB + rm /tmp/tmp.PzkvXvsGNo /tmp/tmp.cP46vE7cCB + return 0 + patch_secret my-cluster-secrets xtrabackup dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=xtrabackup + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.X52PWahHL6 ++ mktemp + local LAST_ERR=/tmp/tmp.M9NTJ2iRNF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.X52PWahHL6 secret/my-cluster-secrets patched + cat /tmp/tmp.M9NTJ2iRNF + rm /tmp/tmp.X52PWahHL6 /tmp/tmp.M9NTJ2iRNF + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oDj7mND9Xa +++ mktemp ++ local LAST_ERR=/tmp/tmp.HxYhXNsSXF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oDj7mND9Xa ++ cat /tmp/tmp.HxYhXNsSXF ++ rm /tmp/tmp.oDj7mND9Xa /tmp/tmp.HxYhXNsSXF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GgKeal3VYk +++ mktemp ++ local LAST_ERR=/tmp/tmp.dgqHh1w4zn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GgKeal3VYk ++ cat /tmp/tmp.dgqHh1w4zn ++ rm /tmp/tmp.GgKeal3VYk /tmp/tmp.dgqHh1w4zn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dUdD5l92fg +++ mktemp ++ local LAST_ERR=/tmp/tmp.K7IeooAdEh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dUdD5l92fg ++ cat /tmp/tmp.K7IeooAdEh ++ rm /tmp/tmp.dUdD5l92fg /tmp/tmp.K7IeooAdEh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B95kO9m3w4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.f2I54S0XYs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B95kO9m3w4 ++ cat /tmp/tmp.f2I54S0XYs ++ rm /tmp/tmp.B95kO9m3w4 /tmp/tmp.f2I54S0XYs ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bSvh9YuMyK +++ mktemp ++ local LAST_ERR=/tmp/tmp.THqAyWdQE5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bSvh9YuMyK ++ cat /tmp/tmp.THqAyWdQE5 ++ rm /tmp/tmp.bSvh9YuMyK /tmp/tmp.THqAyWdQE5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NoQ1SNXN7c +++ mktemp ++ local LAST_ERR=/tmp/tmp.tGFbmOQdRF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NoQ1SNXN7c ++ cat /tmp/tmp.tGFbmOQdRF ++ rm /tmp/tmp.NoQ1SNXN7c /tmp/tmp.tGFbmOQdRF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.L9hWDcZoxe +++ mktemp ++ local LAST_ERR=/tmp/tmp.733TCIagrN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.L9hWDcZoxe ++ cat /tmp/tmp.733TCIagrN ++ rm /tmp/tmp.L9hWDcZoxe /tmp/tmp.733TCIagrN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7Jp7nSyOgn +++ mktemp ++ local LAST_ERR=/tmp/tmp.9jgQba7o4X ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7Jp7nSyOgn ++ cat /tmp/tmp.9jgQba7o4X ++ rm /tmp/tmp.7Jp7nSyOgn /tmp/tmp.9jgQba7o4X ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ir0gbAihSP +++ mktemp ++ local LAST_ERR=/tmp/tmp.tRfLZAeIBV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ir0gbAihSP ++ cat /tmp/tmp.tRfLZAeIBV ++ rm /tmp/tmp.Ir0gbAihSP /tmp/tmp.tRfLZAeIBV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z9OCyKpRUd +++ mktemp ++ local LAST_ERR=/tmp/tmp.aDFeXKzRln ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z9OCyKpRUd ++ cat /tmp/tmp.aDFeXKzRln ++ rm /tmp/tmp.z9OCyKpRUd /tmp/tmp.aDFeXKzRln ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cYn3SSCUzW +++ mktemp ++ local LAST_ERR=/tmp/tmp.S574ShFeLP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cYn3SSCUzW ++ cat /tmp/tmp.S574ShFeLP ++ rm /tmp/tmp.cYn3SSCUzW /tmp/tmp.S574ShFeLP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cDMamBiWZ0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Uz0ANaeybY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cDMamBiWZ0 ++ cat /tmp/tmp.Uz0ANaeybY ++ rm /tmp/tmp.cDMamBiWZ0 /tmp/tmp.Uz0ANaeybY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pYfOWVqaJ7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.AapT72uLTH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pYfOWVqaJ7 ++ cat /tmp/tmp.AapT72uLTH ++ rm /tmp/tmp.pYfOWVqaJ7 /tmp/tmp.AapT72uLTH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1hKUYfSL3F +++ mktemp ++ local LAST_ERR=/tmp/tmp.6pbNpfOw1y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1hKUYfSL3F ++ cat /tmp/tmp.6pbNpfOw1y ++ rm /tmp/tmp.1hKUYfSL3F /tmp/tmp.6pbNpfOw1y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IR5pFRSVMN +++ mktemp ++ local LAST_ERR=/tmp/tmp.tGbh5iZrxN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IR5pFRSVMN ++ cat /tmp/tmp.tGbh5iZrxN ++ rm /tmp/tmp.IR5pFRSVMN /tmp/tmp.tGbh5iZrxN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MT8xoU93Sr +++ mktemp ++ local LAST_ERR=/tmp/tmp.CVpupTChz3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MT8xoU93Sr ++ cat /tmp/tmp.CVpupTChz3 ++ rm /tmp/tmp.MT8xoU93Sr /tmp/tmp.CVpupTChz3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6bRr5Hd0Vm +++ mktemp ++ local LAST_ERR=/tmp/tmp.PRmByvQts8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6bRr5Hd0Vm ++ cat /tmp/tmp.PRmByvQts8 ++ rm /tmp/tmp.6bRr5Hd0Vm /tmp/tmp.PRmByvQts8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qL75kG9veM +++ mktemp ++ local LAST_ERR=/tmp/tmp.zCgtfHKVa5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qL75kG9veM ++ cat /tmp/tmp.zCgtfHKVa5 ++ rm /tmp/tmp.qL75kG9veM /tmp/tmp.zCgtfHKVa5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hZPFwk7gK5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.hgIDXlvLCf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hZPFwk7gK5 ++ cat /tmp/tmp.hgIDXlvLCf ++ rm /tmp/tmp.hZPFwk7gK5 /tmp/tmp.hgIDXlvLCf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m8FbzmHnSQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.yL6AFTdKC4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.m8FbzmHnSQ ++ cat /tmp/tmp.yL6AFTdKC4 ++ rm /tmp/tmp.m8FbzmHnSQ /tmp/tmp.yL6AFTdKC4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wBJxW2vmAd +++ mktemp ++ local LAST_ERR=/tmp/tmp.iUU3gYTgkh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wBJxW2vmAd ++ cat /tmp/tmp.iUU3gYTgkh ++ rm /tmp/tmp.wBJxW2vmAd /tmp/tmp.iUU3gYTgkh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rv5tAE1ioO +++ mktemp ++ local LAST_ERR=/tmp/tmp.AQ9KO6CCf9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rv5tAE1ioO ++ cat /tmp/tmp.AQ9KO6CCf9 ++ rm /tmp/tmp.rv5tAE1ioO /tmp/tmp.AQ9KO6CCf9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U3E1tto6Cx +++ mktemp ++ local LAST_ERR=/tmp/tmp.cAChFgnJ3z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.U3E1tto6Cx ++ cat /tmp/tmp.cAChFgnJ3z ++ rm /tmp/tmp.U3E1tto6Cx /tmp/tmp.cAChFgnJ3z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Kcp61a5bkd +++ mktemp ++ local LAST_ERR=/tmp/tmp.IgIVVgf7Q7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Kcp61a5bkd ++ cat /tmp/tmp.IgIVVgf7Q7 ++ rm /tmp/tmp.Kcp61a5bkd /tmp/tmp.IgIVVgf7Q7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aDGIz05BzI +++ mktemp ++ local LAST_ERR=/tmp/tmp.GT6edCw0j5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aDGIz05BzI ++ cat /tmp/tmp.GT6edCw0j5 ++ rm /tmp/tmp.aDGIz05BzI /tmp/tmp.GT6edCw0j5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zkZEGtlUUF +++ mktemp ++ local LAST_ERR=/tmp/tmp.taNqMKnG2q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zkZEGtlUUF ++ cat /tmp/tmp.taNqMKnG2q ++ rm /tmp/tmp.zkZEGtlUUF /tmp/tmp.taNqMKnG2q ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KHiZZwOwGr +++ mktemp ++ local LAST_ERR=/tmp/tmp.2ybujezGpZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KHiZZwOwGr ++ cat /tmp/tmp.2ybujezGpZ ++ rm /tmp/tmp.KHiZZwOwGr /tmp/tmp.2ybujezGpZ ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.J77HA5OvvP +++ mktemp ++ local LAST_ERR=/tmp/tmp.Urxvzbcphp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.J77HA5OvvP ++ cat /tmp/tmp.Urxvzbcphp ++ rm /tmp/tmp.J77HA5OvvP /tmp/tmp.Urxvzbcphp ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.8ZVAO3852y ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.J0krYCTQDh +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.8ZVAO3852y +++++ cat /tmp/tmp.J0krYCTQDh +++++ rm /tmp/tmp.8ZVAO3852y /tmp/tmp.J0krYCTQDh +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.CrfVZvnZkP ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.iSWm0JuWhq +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.CrfVZvnZkP +++++ cat /tmp/tmp.iSWm0JuWhq +++++ rm /tmp/tmp.CrfVZvnZkP /tmp/tmp.iSWm0JuWhq +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tvuWj0cNgL +++ mktemp ++ local LAST_ERR=/tmp/tmp.c583R590fp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tvuWj0cNgL ++ cat /tmp/tmp.c583R590fp ++ rm /tmp/tmp.tvuWj0cNgL /tmp/tmp.c583R590fp ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd_local select-3 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 '' pxc + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-3-84.sql ']' + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 pxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + '[' '!' -s /tmp/tmp.QDkOK88Jf9/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-3.sql /tmp/tmp.QDkOK88Jf9/select-3.sql + desc 'test monitor' + set +o xtrace ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.pgtHKyIFu7 ++ mktemp + local LAST_ERR=/tmp/tmp.eHsQgRH2Gi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pgtHKyIFu7 secret/my-cluster-secrets patched + cat /tmp/tmp.eHsQgRH2Gi + rm /tmp/tmp.pgtHKyIFu7 /tmp/tmp.eHsQgRH2Gi + return 0 + wait_for_password_propagation my-cluster-secrets monitor some-name + local secret=my-cluster-secrets + local user=monitor + local cluster=some-name + local max_retry=600 + local root_pass + local retry ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.nOLkCMTisT +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZhV9xFWRHP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nOLkCMTisT ++ cat /tmp/tmp.ZhV9xFWRHP ++ rm /tmp/tmp.nOLkCMTisT /tmp/tmp.ZhV9xFWRHP ++ return 0 + root_pass=test-password + [[ -z test-password ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + retry=0 + is_password_updated monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + grep additional_password + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KGpohQFD4U +++ mktemp ++ local LAST_ERR=/tmp/tmp.eNcvVODCXI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KGpohQFD4U ++ cat /tmp/tmp.eNcvVODCXI ++ rm /tmp/tmp.KGpohQFD4U /tmp/tmp.eNcvVODCXI ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} + sleep 5 + retry=0 + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=1 + [[ 1 -eq 0 ]] + [[ 1 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=2 + [[ 2 -eq 0 ]] + [[ 2 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=3 + [[ 3 -eq 0 ]] + [[ 3 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=4 + [[ 4 -eq 0 ]] + [[ 4 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=5 + [[ 5 -eq 0 ]] + [[ 5 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=6 + [[ 6 -eq 0 ]] + [[ 6 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=7 + [[ 7 -eq 0 ]] + [[ 7 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=8 + [[ 8 -eq 0 ]] + [[ 8 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"}' + [[ {"additional_password": "$A$005$_7\\n_u\\u0001\\u0001e\\u0019\\u0004!\\\\z>jEKPj3pt4DlspzaiG7gwbzhO3IjcLKcHkhDBg0BaiTDzo7Y7"} =~ additional_password ]] + return 1 + sleep 1 + retry=9 + [[ 9 -eq 0 ]] + [[ 9 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result=NULL + [[ NULL =~ additional_password ]] + [[ -z NULL ]] + [[ NULL =~ ^[[:space:]]*$ ]] + [[ NULL =~ NULL ]] + return 0 + echo 'Old password successfully discarded after 9 retries' Old password successfully discarded after 9 retries + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c8V9oOD2rm +++ mktemp ++ local LAST_ERR=/tmp/tmp.RGDKge0nE0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.c8V9oOD2rm ++ cat /tmp/tmp.RGDKge0nE0 ++ rm /tmp/tmp.c8V9oOD2rm /tmp/tmp.RGDKge0nE0 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f7w7y0Wmso +++ mktemp ++ local LAST_ERR=/tmp/tmp.ukNYqKCRUy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f7w7y0Wmso ++ cat /tmp/tmp.ukNYqKCRUy ++ rm /tmp/tmp.f7w7y0Wmso /tmp/tmp.ukNYqKCRUy ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.8hYLsw6G7V ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.PAPOAInETv +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.8hYLsw6G7V +++++ cat /tmp/tmp.PAPOAInETv +++++ rm /tmp/tmp.8hYLsw6G7V /tmp/tmp.PAPOAInETv +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.s0jmjhUBDm ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.pVRKLw9NL5 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.s0jmjhUBDm +++++ cat /tmp/tmp.pVRKLw9NL5 +++++ rm /tmp/tmp.s0jmjhUBDm /tmp/tmp.pVRKLw9NL5 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YNs5WYu8PY +++ mktemp ++ local LAST_ERR=/tmp/tmp.xUhm110wKV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YNs5WYu8PY ++ cat /tmp/tmp.xUhm110wKV ++ rm /tmp/tmp.YNs5WYu8PY /tmp/tmp.xUhm110wKV ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.isCWjW40Sv +++ mktemp ++ local LAST_ERR=/tmp/tmp.qdvvZoUWJH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.isCWjW40Sv ++ cat /tmp/tmp.qdvvZoUWJH ++ rm /tmp/tmp.isCWjW40Sv /tmp/tmp.qdvvZoUWJH ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + compare_mysql_cmd_local proxysql-cfg 'SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=proxysql-cfg + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/proxysql-cfg.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/proxysql-cfg-84.sql ']' + run_mysql_local 'SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.QDkOK88Jf9/proxysql-cfg.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/proxysql-cfg.sql /tmp/tmp.QDkOK88Jf9/proxysql-cfg.sql + desc 'test operator' + set +o xtrace ----------------------------------------------------------------------------------- test operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets operator dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=operator + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.bNO4YN3duf ++ mktemp + local LAST_ERR=/tmp/tmp.yf2Iwohq1a + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.bNO4YN3duf secret/my-cluster-secrets patched + cat /tmp/tmp.yf2Iwohq1a + rm /tmp/tmp.bNO4YN3duf /tmp/tmp.yf2Iwohq1a + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VXQtymID16 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qIIop3cQwY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VXQtymID16 ++ cat /tmp/tmp.qIIop3cQwY ++ rm /tmp/tmp.VXQtymID16 /tmp/tmp.qIIop3cQwY ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e9O3LSjley +++ mktemp ++ local LAST_ERR=/tmp/tmp.5M4GWF0ZA6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.e9O3LSjley ++ cat /tmp/tmp.5M4GWF0ZA6 ++ rm /tmp/tmp.e9O3LSjley /tmp/tmp.5M4GWF0ZA6 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.2ogbtaVrXy ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.qbzbB5NQJg +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.2ogbtaVrXy +++++ cat /tmp/tmp.qbzbB5NQJg +++++ rm /tmp/tmp.2ogbtaVrXy /tmp/tmp.qbzbB5NQJg +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.0m09ZEiAlt ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.BfZsq1ejzb +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.0m09ZEiAlt +++++ cat /tmp/tmp.BfZsq1ejzb +++++ rm /tmp/tmp.0m09ZEiAlt /tmp/tmp.BfZsq1ejzb +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BejbsqA6u5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Zhc1pgb0O3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BejbsqA6u5 ++ cat /tmp/tmp.Zhc1pgb0O3 ++ rm /tmp/tmp.BejbsqA6u5 /tmp/tmp.Zhc1pgb0O3 ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vUN7Zlal1i +++ mktemp ++ local LAST_ERR=/tmp/tmp.rNx4lxNMga ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vUN7Zlal1i ++ cat /tmp/tmp.rNx4lxNMga ++ rm /tmp/tmp.vUN7Zlal1i /tmp/tmp.rNx4lxNMga ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + desc 'change secret name' + set +o xtrace ----------------------------------------------------------------------------------- change secret name ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.s3TQSm83Hu ++ mktemp + local LAST_ERR=/tmp/tmp.DLQLOBS7Dr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.s3TQSm83Hu perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.DLQLOBS7Dr + rm /tmp/tmp.s3TQSm83Hu /tmp/tmp.DLQLOBS7Dr + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vT6p76hb3D +++ mktemp ++ local LAST_ERR=/tmp/tmp.eUneGWeGfg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vT6p76hb3D ++ cat /tmp/tmp.eUneGWeGfg ++ rm /tmp/tmp.vT6p76hb3D /tmp/tmp.eUneGWeGfg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.60KZeVkamw +++ mktemp ++ local LAST_ERR=/tmp/tmp.20wVNTfhyh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.60KZeVkamw ++ cat /tmp/tmp.20wVNTfhyh ++ rm /tmp/tmp.60KZeVkamw /tmp/tmp.20wVNTfhyh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.snL91GWuOJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.bDTnRNUIfr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.snL91GWuOJ ++ cat /tmp/tmp.bDTnRNUIfr ++ rm /tmp/tmp.snL91GWuOJ /tmp/tmp.bDTnRNUIfr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PgfoBlUEGq +++ mktemp ++ local LAST_ERR=/tmp/tmp.K7qlLpBRge ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PgfoBlUEGq ++ cat /tmp/tmp.K7qlLpBRge ++ rm /tmp/tmp.PgfoBlUEGq /tmp/tmp.K7qlLpBRge ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DAH6p4A98v +++ mktemp ++ local LAST_ERR=/tmp/tmp.h29rLb7N5f ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DAH6p4A98v ++ cat /tmp/tmp.h29rLb7N5f ++ rm /tmp/tmp.DAH6p4A98v /tmp/tmp.h29rLb7N5f ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7wheanHdkM +++ mktemp ++ local LAST_ERR=/tmp/tmp.wursf3TFKz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7wheanHdkM ++ cat /tmp/tmp.wursf3TFKz ++ rm /tmp/tmp.7wheanHdkM /tmp/tmp.wursf3TFKz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I5TI1ET2Vu +++ mktemp ++ local LAST_ERR=/tmp/tmp.dU2ZdY26rY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.I5TI1ET2Vu ++ cat /tmp/tmp.dU2ZdY26rY ++ rm /tmp/tmp.I5TI1ET2Vu /tmp/tmp.dU2ZdY26rY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KjjKWkwSh8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9yC1rpunuC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KjjKWkwSh8 ++ cat /tmp/tmp.9yC1rpunuC ++ rm /tmp/tmp.KjjKWkwSh8 /tmp/tmp.9yC1rpunuC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Iqvs7FPl59 +++ mktemp ++ local LAST_ERR=/tmp/tmp.TvR2Z5f68q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Iqvs7FPl59 ++ cat /tmp/tmp.TvR2Z5f68q ++ rm /tmp/tmp.Iqvs7FPl59 /tmp/tmp.TvR2Z5f68q ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o6PeWuCro9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zOMLAAK43m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o6PeWuCro9 ++ cat /tmp/tmp.zOMLAAK43m ++ rm /tmp/tmp.o6PeWuCro9 /tmp/tmp.zOMLAAK43m ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LFdveQqIix +++ mktemp ++ local LAST_ERR=/tmp/tmp.f3hg4vBkeq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LFdveQqIix ++ cat /tmp/tmp.f3hg4vBkeq ++ rm /tmp/tmp.LFdveQqIix /tmp/tmp.f3hg4vBkeq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ov3togqIns +++ mktemp ++ local LAST_ERR=/tmp/tmp.PqIqZxN74I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ov3togqIns ++ cat /tmp/tmp.PqIqZxN74I ++ rm /tmp/tmp.Ov3togqIns /tmp/tmp.PqIqZxN74I ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e7q3R44zyI +++ mktemp ++ local LAST_ERR=/tmp/tmp.uFnrdrmEUZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.e7q3R44zyI ++ cat /tmp/tmp.uFnrdrmEUZ ++ rm /tmp/tmp.e7q3R44zyI /tmp/tmp.uFnrdrmEUZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KLqfC043hf +++ mktemp ++ local LAST_ERR=/tmp/tmp.kgcE2epf8p ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KLqfC043hf ++ cat /tmp/tmp.kgcE2epf8p ++ rm /tmp/tmp.KLqfC043hf /tmp/tmp.kgcE2epf8p ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xwiaRiIZzl +++ mktemp ++ local LAST_ERR=/tmp/tmp.gyEfEEAZhV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xwiaRiIZzl ++ cat /tmp/tmp.gyEfEEAZhV ++ rm /tmp/tmp.xwiaRiIZzl /tmp/tmp.gyEfEEAZhV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EO4X0O5Maq +++ mktemp ++ local LAST_ERR=/tmp/tmp.COrWSMdNp3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EO4X0O5Maq ++ cat /tmp/tmp.COrWSMdNp3 ++ rm /tmp/tmp.EO4X0O5Maq /tmp/tmp.COrWSMdNp3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fR41DKgFa3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ey2A1HShJ4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fR41DKgFa3 ++ cat /tmp/tmp.Ey2A1HShJ4 ++ rm /tmp/tmp.fR41DKgFa3 /tmp/tmp.Ey2A1HShJ4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rJj0l38AlD +++ mktemp ++ local LAST_ERR=/tmp/tmp.iRrWLHLHqQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rJj0l38AlD ++ cat /tmp/tmp.iRrWLHLHqQ ++ rm /tmp/tmp.rJj0l38AlD /tmp/tmp.iRrWLHLHqQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A0pqsqSMXW +++ mktemp ++ local LAST_ERR=/tmp/tmp.8YjnT4Dab1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.A0pqsqSMXW ++ cat /tmp/tmp.8YjnT4Dab1 ++ rm /tmp/tmp.A0pqsqSMXW /tmp/tmp.8YjnT4Dab1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eAlztEyzqi +++ mktemp ++ local LAST_ERR=/tmp/tmp.AXDyPmJ5qZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eAlztEyzqi ++ cat /tmp/tmp.AXDyPmJ5qZ ++ rm /tmp/tmp.eAlztEyzqi /tmp/tmp.AXDyPmJ5qZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HP8sraES2l +++ mktemp ++ local LAST_ERR=/tmp/tmp.yaxCaiKbio ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HP8sraES2l ++ cat /tmp/tmp.yaxCaiKbio ++ rm /tmp/tmp.HP8sraES2l /tmp/tmp.yaxCaiKbio ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WEYtZeXx0k +++ mktemp ++ local LAST_ERR=/tmp/tmp.AV0Qi7faKS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WEYtZeXx0k ++ cat /tmp/tmp.AV0Qi7faKS ++ rm /tmp/tmp.WEYtZeXx0k /tmp/tmp.AV0Qi7faKS ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pxOnt1vBEQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.UhMNXoqsgt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pxOnt1vBEQ ++ cat /tmp/tmp.UhMNXoqsgt ++ rm /tmp/tmp.pxOnt1vBEQ /tmp/tmp.UhMNXoqsgt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MsjqaKmW2o +++ mktemp ++ local LAST_ERR=/tmp/tmp.5KXbECZjzA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MsjqaKmW2o ++ cat /tmp/tmp.5KXbECZjzA ++ rm /tmp/tmp.MsjqaKmW2o /tmp/tmp.5KXbECZjzA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qwAMUHi9cw +++ mktemp ++ local LAST_ERR=/tmp/tmp.XPF4s7b7Xu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qwAMUHi9cw ++ cat /tmp/tmp.XPF4s7b7Xu ++ rm /tmp/tmp.qwAMUHi9cw /tmp/tmp.XPF4s7b7Xu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LpO1sVLvjx +++ mktemp ++ local LAST_ERR=/tmp/tmp.RZb8Zl3PmB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LpO1sVLvjx ++ cat /tmp/tmp.RZb8Zl3PmB ++ rm /tmp/tmp.LpO1sVLvjx /tmp/tmp.RZb8Zl3PmB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jgPdE3JwMA +++ mktemp ++ local LAST_ERR=/tmp/tmp.xO18orH9oU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jgPdE3JwMA ++ cat /tmp/tmp.xO18orH9oU ++ rm /tmp/tmp.jgPdE3JwMA /tmp/tmp.xO18orH9oU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 26 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jq2ppoaumW +++ mktemp ++ local LAST_ERR=/tmp/tmp.FvqdtgcTjP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jq2ppoaumW ++ cat /tmp/tmp.FvqdtgcTjP ++ rm /tmp/tmp.jq2ppoaumW /tmp/tmp.FvqdtgcTjP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 27 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UeBslBZttg +++ mktemp ++ local LAST_ERR=/tmp/tmp.IRW5Dz3Trq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UeBslBZttg ++ cat /tmp/tmp.IRW5Dz3Trq ++ rm /tmp/tmp.UeBslBZttg /tmp/tmp.IRW5Dz3Trq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 28 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cM9DwFDSE4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yZF4dMyBbg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cM9DwFDSE4 ++ cat /tmp/tmp.yZF4dMyBbg ++ rm /tmp/tmp.cM9DwFDSE4 /tmp/tmp.yZF4dMyBbg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 29 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2kJg3vrkPm +++ mktemp ++ local LAST_ERR=/tmp/tmp.ozxSuMNYzt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2kJg3vrkPm ++ cat /tmp/tmp.ozxSuMNYzt ++ rm /tmp/tmp.2kJg3vrkPm /tmp/tmp.ozxSuMNYzt ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GwkESynVRd +++ mktemp ++ local LAST_ERR=/tmp/tmp.C3WybPDdju ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GwkESynVRd ++ cat /tmp/tmp.C3WybPDdju ++ rm /tmp/tmp.GwkESynVRd /tmp/tmp.C3WybPDdju ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.X1yOQH30Wp ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.hHxmknBMsz +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.X1yOQH30Wp +++++ cat /tmp/tmp.hHxmknBMsz +++++ rm /tmp/tmp.X1yOQH30Wp /tmp/tmp.hHxmknBMsz +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.MM83Cv8QWm ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.jm7Ycp0FPf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.MM83Cv8QWm +++++ cat /tmp/tmp.jm7Ycp0FPf +++++ rm /tmp/tmp.MM83Cv8QWm /tmp/tmp.jm7Ycp0FPf +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AXs6LJGatA +++ mktemp ++ local LAST_ERR=/tmp/tmp.iLwdkNnIDO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AXs6LJGatA ++ cat /tmp/tmp.iLwdkNnIDO ++ rm /tmp/tmp.AXs6LJGatA /tmp/tmp.iLwdkNnIDO ++ return 0 + [[ 2 == \2 ]] + echo + desc 'test new operator' + set +o xtrace ----------------------------------------------------------------------------------- test new operator ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + patch_secret my-cluster-secrets-2 operator dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets-2 + local key=operator + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.ooDUO8HRfG ++ mktemp + local LAST_ERR=/tmp/tmp.2UMlvtls85 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ooDUO8HRfG secret/my-cluster-secrets-2 patched + cat /tmp/tmp.2UMlvtls85 + rm /tmp/tmp.ooDUO8HRfG /tmp/tmp.2UMlvtls85 + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VZPIiLyyBI +++ mktemp ++ local LAST_ERR=/tmp/tmp.RpEfuGBJRo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VZPIiLyyBI ++ cat /tmp/tmp.RpEfuGBJRo ++ rm /tmp/tmp.VZPIiLyyBI /tmp/tmp.RpEfuGBJRo ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZYi4VingTM +++ mktemp ++ local LAST_ERR=/tmp/tmp.KFXnzpQKeS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZYi4VingTM ++ cat /tmp/tmp.KFXnzpQKeS ++ rm /tmp/tmp.ZYi4VingTM /tmp/tmp.KFXnzpQKeS ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cjTIdUxufN +++ mktemp ++ local LAST_ERR=/tmp/tmp.IY8dfgqJxO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cjTIdUxufN ++ cat /tmp/tmp.IY8dfgqJxO ++ rm /tmp/tmp.cjTIdUxufN /tmp/tmp.IY8dfgqJxO ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.XtNFyjIM9n ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.C3OaJs0hir +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.XtNFyjIM9n +++++ cat /tmp/tmp.C3OaJs0hir +++++ rm /tmp/tmp.XtNFyjIM9n /tmp/tmp.C3OaJs0hir +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.VFQB5mRrf6 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.wOItfZnT5N +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.VFQB5mRrf6 +++++ cat /tmp/tmp.wOItfZnT5N +++++ rm /tmp/tmp.VFQB5mRrf6 /tmp/tmp.wOItfZnT5N +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.J2SvXBLmc0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.IEZbNxAGq1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.J2SvXBLmc0 ++ cat /tmp/tmp.IEZbNxAGq1 ++ rm /tmp/tmp.J2SvXBLmc0 /tmp/tmp.IEZbNxAGq1 ++ return 0 + [[ 2 == \2 ]] + echo + sleep 20 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U3tfkHGvdy +++ mktemp ++ local LAST_ERR=/tmp/tmp.hVVm5EbDG6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.U3tfkHGvdy ++ cat /tmp/tmp.hVVm5EbDG6 ++ rm /tmp/tmp.U3tfkHGvdy /tmp/tmp.hVVm5EbDG6 ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql ++ getSecretData my-cluster-secrets-2 root ++ local secretName=my-cluster-secrets-2 ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.GCu9xGMeJr +++ mktemp ++ local LAST_ERR=/tmp/tmp.U4y7X2puUX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GCu9xGMeJr ++ cat /tmp/tmp.U4y7X2puUX ++ rm /tmp/tmp.GCu9xGMeJr /tmp/tmp.U4y7X2puUX ++ return 0 + newpass='2[ElArS&%&48fk01b>' + desc 'test new users sync' + set +o xtrace ----------------------------------------------------------------------------------- test new users sync ----------------------------------------------------------------------------------- + run_mysql 'CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''2[ElArS&%&48fk01b>'\'';' '-h some-name-pxc -uroot -p'\''2[ElArS&%&48fk01b>'\''' + local 'command=CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''2[ElArS&%&48fk01b>'\'';' + local 'uri=-h some-name-pxc -uroot -p'\''2[ElArS&%&48fk01b>'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dfTw4SYvsD +++ mktemp ++ local LAST_ERR=/tmp/tmp.H3CJmSkvZ6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dfTw4SYvsD ++ cat /tmp/tmp.H3CJmSkvZ6 ++ rm /tmp/tmp.dfTw4SYvsD /tmp/tmp.H3CJmSkvZ6 ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 80 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''2[ElArS&%&48fk01b>'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''2[ElArS&%&48fk01b>'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''2[ElArS&%&48fk01b>'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''2[ElArS&%&48fk01b>'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gSxY4bZyMU +++ mktemp ++ local LAST_ERR=/tmp/tmp.wAtMqRNAj3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gSxY4bZyMU ++ cat /tmp/tmp.wAtMqRNAj3 ++ rm /tmp/tmp.gSxY4bZyMU /tmp/tmp.wAtMqRNAj3 ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.EvogazPSKB +++ mktemp ++ local LAST_ERR=/tmp/tmp.8FWHt6VN4j ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EvogazPSKB ++ cat /tmp/tmp.8FWHt6VN4j ++ rm /tmp/tmp.EvogazPSKB /tmp/tmp.8FWHt6VN4j ++ return 0 + pass=test-password2 + desc 'check secret without operator' + set +o xtrace ----------------------------------------------------------------------------------- check secret without operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.J6TnwwIl9u ++ mktemp + local LAST_ERR=/tmp/tmp.WmhpPcYyEg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.J6TnwwIl9u secret/my-cluster-secrets-2 configured + cat /tmp/tmp.WmhpPcYyEg Warning: resource secrets/my-cluster-secrets-2 is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.J6TnwwIl9u /tmp/tmp.WmhpPcYyEg + return 0 + sleep 60 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f96mp9iNMk +++ mktemp ++ local LAST_ERR=/tmp/tmp.SOvlgMItg2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f96mp9iNMk ++ cat /tmp/tmp.SOvlgMItg2 ++ rm /tmp/tmp.f96mp9iNMk /tmp/tmp.SOvlgMItg2 ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.QDkOK88Jf9/select-4.sql + desc 'test enable haproxy and disable proxysql ' + set +o xtrace ----------------------------------------------------------------------------------- test enable haproxy and disable proxysql ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch pxc some-name --type=merge -p '{"spec":{"haproxy":{"enabled":true},"proxysql":{"enabled":false},"secretsName":"my-cluster-secrets"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.Yosh0tyMXv ++ mktemp + local LAST_ERR=/tmp/tmp.1WxMciTCOx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge -p '{"spec":{"haproxy":{"enabled":true},"proxysql":{"enabled":false},"secretsName":"my-cluster-secrets"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Yosh0tyMXv perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.1WxMciTCOx + rm /tmp/tmp.Yosh0tyMXv /tmp/tmp.1WxMciTCOx + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4ok3WVctzf +++ mktemp ++ local LAST_ERR=/tmp/tmp.yjIS5xMlyQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4ok3WVctzf ++ cat /tmp/tmp.yjIS5xMlyQ ++ rm /tmp/tmp.4ok3WVctzf /tmp/tmp.yjIS5xMlyQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3WU8UHDCtr +++ mktemp ++ local LAST_ERR=/tmp/tmp.sgPnid5czP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3WU8UHDCtr ++ cat /tmp/tmp.sgPnid5czP ++ rm /tmp/tmp.3WU8UHDCtr /tmp/tmp.sgPnid5czP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tFdOtyndRI +++ mktemp ++ local LAST_ERR=/tmp/tmp.tPg1Vm0DZG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tFdOtyndRI ++ cat /tmp/tmp.tPg1Vm0DZG ++ rm /tmp/tmp.tFdOtyndRI /tmp/tmp.tPg1Vm0DZG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TrR3y2qgp2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.5HnFsVkQ9a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TrR3y2qgp2 ++ cat /tmp/tmp.5HnFsVkQ9a ++ rm /tmp/tmp.TrR3y2qgp2 /tmp/tmp.5HnFsVkQ9a ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YDZI0vsGnk +++ mktemp ++ local LAST_ERR=/tmp/tmp.GaiAYoJ7yv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YDZI0vsGnk ++ cat /tmp/tmp.GaiAYoJ7yv ++ rm /tmp/tmp.YDZI0vsGnk /tmp/tmp.GaiAYoJ7yv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1Vh4MjJ5F1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PzEGC3lytV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1Vh4MjJ5F1 ++ cat /tmp/tmp.PzEGC3lytV ++ rm /tmp/tmp.1Vh4MjJ5F1 /tmp/tmp.PzEGC3lytV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rxUpWAygMa +++ mktemp ++ local LAST_ERR=/tmp/tmp.2GWrsThpdV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rxUpWAygMa ++ cat /tmp/tmp.2GWrsThpdV ++ rm /tmp/tmp.rxUpWAygMa /tmp/tmp.2GWrsThpdV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fdfqsctIp0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ufqhle3Vt3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fdfqsctIp0 ++ cat /tmp/tmp.Ufqhle3Vt3 ++ rm /tmp/tmp.fdfqsctIp0 /tmp/tmp.Ufqhle3Vt3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eMAafJotpn +++ mktemp ++ local LAST_ERR=/tmp/tmp.oakRkXUYni ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eMAafJotpn ++ cat /tmp/tmp.oakRkXUYni ++ rm /tmp/tmp.eMAafJotpn /tmp/tmp.oakRkXUYni ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0UxoUUNDas +++ mktemp ++ local LAST_ERR=/tmp/tmp.OZYMY3yLx8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0UxoUUNDas ++ cat /tmp/tmp.OZYMY3yLx8 ++ rm /tmp/tmp.0UxoUUNDas /tmp/tmp.OZYMY3yLx8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kcJUY2z4Ls +++ mktemp ++ local LAST_ERR=/tmp/tmp.SjcxKKcWjj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kcJUY2z4Ls ++ cat /tmp/tmp.SjcxKKcWjj ++ rm /tmp/tmp.kcJUY2z4Ls /tmp/tmp.SjcxKKcWjj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8pJ0ZrdEXM +++ mktemp ++ local LAST_ERR=/tmp/tmp.M7MqbNlmRm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8pJ0ZrdEXM ++ cat /tmp/tmp.M7MqbNlmRm ++ rm /tmp/tmp.8pJ0ZrdEXM /tmp/tmp.M7MqbNlmRm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Nx7sI2Ft8m +++ mktemp ++ local LAST_ERR=/tmp/tmp.6MeatOKYgN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Nx7sI2Ft8m ++ cat /tmp/tmp.6MeatOKYgN ++ rm /tmp/tmp.Nx7sI2Ft8m /tmp/tmp.6MeatOKYgN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.knOieB4pi7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PXFlNqDRQf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.knOieB4pi7 ++ cat /tmp/tmp.PXFlNqDRQf ++ rm /tmp/tmp.knOieB4pi7 /tmp/tmp.PXFlNqDRQf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6r7WI4LV4L +++ mktemp ++ local LAST_ERR=/tmp/tmp.50C8jYPws6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6r7WI4LV4L ++ cat /tmp/tmp.50C8jYPws6 ++ rm /tmp/tmp.6r7WI4LV4L /tmp/tmp.50C8jYPws6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yk5nqNH46N +++ mktemp ++ local LAST_ERR=/tmp/tmp.H9bMmcMOHZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yk5nqNH46N ++ cat /tmp/tmp.H9bMmcMOHZ ++ rm /tmp/tmp.yk5nqNH46N /tmp/tmp.H9bMmcMOHZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tK1x3VjKHt +++ mktemp ++ local LAST_ERR=/tmp/tmp.rXIdt92Cif ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tK1x3VjKHt ++ cat /tmp/tmp.rXIdt92Cif ++ rm /tmp/tmp.tK1x3VjKHt /tmp/tmp.rXIdt92Cif ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cF92EDJXkw +++ mktemp ++ local LAST_ERR=/tmp/tmp.I3JI1rVSMG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cF92EDJXkw ++ cat /tmp/tmp.I3JI1rVSMG ++ rm /tmp/tmp.cF92EDJXkw /tmp/tmp.I3JI1rVSMG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pjCRV2LK1V +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ev8EkUSW1f ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pjCRV2LK1V ++ cat /tmp/tmp.Ev8EkUSW1f ++ rm /tmp/tmp.pjCRV2LK1V /tmp/tmp.Ev8EkUSW1f ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hCpEMY1l4T +++ mktemp ++ local LAST_ERR=/tmp/tmp.k260R9JeKK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hCpEMY1l4T ++ cat /tmp/tmp.k260R9JeKK ++ rm /tmp/tmp.hCpEMY1l4T /tmp/tmp.k260R9JeKK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MVBnPAbjS2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tqtvPOwGta ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MVBnPAbjS2 ++ cat /tmp/tmp.tqtvPOwGta ++ rm /tmp/tmp.MVBnPAbjS2 /tmp/tmp.tqtvPOwGta ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zk75oX33Yt +++ mktemp ++ local LAST_ERR=/tmp/tmp.qomVclYFGO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zk75oX33Yt ++ cat /tmp/tmp.qomVclYFGO ++ rm /tmp/tmp.zk75oX33Yt /tmp/tmp.qomVclYFGO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fhB60m83my +++ mktemp ++ local LAST_ERR=/tmp/tmp.c17FZ1oHpx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fhB60m83my ++ cat /tmp/tmp.c17FZ1oHpx ++ rm /tmp/tmp.fhB60m83my /tmp/tmp.c17FZ1oHpx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gK3gTq3qBF +++ mktemp ++ local LAST_ERR=/tmp/tmp.s8hURyfu7Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gK3gTq3qBF ++ cat /tmp/tmp.s8hURyfu7Y ++ rm /tmp/tmp.gK3gTq3qBF /tmp/tmp.s8hURyfu7Y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9ZLdc4LHQD +++ mktemp ++ local LAST_ERR=/tmp/tmp.dSc7Osk04x ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9ZLdc4LHQD ++ cat /tmp/tmp.dSc7Osk04x ++ rm /tmp/tmp.9ZLdc4LHQD /tmp/tmp.dSc7Osk04x ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DWoeOhKAzR +++ mktemp ++ local LAST_ERR=/tmp/tmp.JeFmrGVBnP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DWoeOhKAzR ++ cat /tmp/tmp.JeFmrGVBnP ++ rm /tmp/tmp.DWoeOhKAzR /tmp/tmp.JeFmrGVBnP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ONQWIYSznd +++ mktemp ++ local LAST_ERR=/tmp/tmp.NW3rrAIuK2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ONQWIYSznd ++ cat /tmp/tmp.NW3rrAIuK2 ++ rm /tmp/tmp.ONQWIYSznd /tmp/tmp.NW3rrAIuK2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 26 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cRWLtg1QHt +++ mktemp ++ local LAST_ERR=/tmp/tmp.TcvIaZUllA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cRWLtg1QHt ++ cat /tmp/tmp.TcvIaZUllA ++ rm /tmp/tmp.cRWLtg1QHt /tmp/tmp.TcvIaZUllA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 27 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2LZxmy2RWw +++ mktemp ++ local LAST_ERR=/tmp/tmp.uTxx7zKgmW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2LZxmy2RWw ++ cat /tmp/tmp.uTxx7zKgmW ++ rm /tmp/tmp.2LZxmy2RWw /tmp/tmp.uTxx7zKgmW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 28 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zM9dHKl9og +++ mktemp ++ local LAST_ERR=/tmp/tmp.YgodupuL1K ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zM9dHKl9og ++ cat /tmp/tmp.YgodupuL1K ++ rm /tmp/tmp.zM9dHKl9og /tmp/tmp.YgodupuL1K ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 29 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4uYWHydYpT +++ mktemp ++ local LAST_ERR=/tmp/tmp.EUgUk1ogDv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4uYWHydYpT ++ cat /tmp/tmp.EUgUk1ogDv ++ rm /tmp/tmp.4uYWHydYpT /tmp/tmp.EUgUk1ogDv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 30 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4dzMTmLTZB +++ mktemp ++ local LAST_ERR=/tmp/tmp.3CV10iiyFx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4dzMTmLTZB ++ cat /tmp/tmp.3CV10iiyFx ++ rm /tmp/tmp.4dzMTmLTZB /tmp/tmp.3CV10iiyFx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 31 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mMx4AFVTE4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.2xbOtx4hqO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mMx4AFVTE4 ++ cat /tmp/tmp.2xbOtx4hqO ++ rm /tmp/tmp.mMx4AFVTE4 /tmp/tmp.2xbOtx4hqO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 32 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TTIArkfCib +++ mktemp ++ local LAST_ERR=/tmp/tmp.FCou2LutPD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TTIArkfCib ++ cat /tmp/tmp.FCou2LutPD ++ rm /tmp/tmp.TTIArkfCib /tmp/tmp.FCou2LutPD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 33 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jxYpLFqkI8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bBj3KRVAkE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jxYpLFqkI8 ++ cat /tmp/tmp.bBj3KRVAkE ++ rm /tmp/tmp.jxYpLFqkI8 /tmp/tmp.bBj3KRVAkE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 34 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U3mKmPuivT +++ mktemp ++ local LAST_ERR=/tmp/tmp.M5At4eAJvA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.U3mKmPuivT ++ cat /tmp/tmp.M5At4eAJvA ++ rm /tmp/tmp.U3mKmPuivT /tmp/tmp.M5At4eAJvA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 35 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZhB2rzQtOA +++ mktemp ++ local LAST_ERR=/tmp/tmp.GEvQlZ9a8a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZhB2rzQtOA ++ cat /tmp/tmp.GEvQlZ9a8a ++ rm /tmp/tmp.ZhB2rzQtOA /tmp/tmp.GEvQlZ9a8a ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 36 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.W4Ray62cDI +++ mktemp ++ local LAST_ERR=/tmp/tmp.IepS0SdsEt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.W4Ray62cDI ++ cat /tmp/tmp.IepS0SdsEt ++ rm /tmp/tmp.W4Ray62cDI /tmp/tmp.IepS0SdsEt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 37 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x5QnhFYTxP +++ mktemp ++ local LAST_ERR=/tmp/tmp.83oRSbrx2Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.x5QnhFYTxP ++ cat /tmp/tmp.83oRSbrx2Z ++ rm /tmp/tmp.x5QnhFYTxP /tmp/tmp.83oRSbrx2Z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 38 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kdjwuqdK1R +++ mktemp ++ local LAST_ERR=/tmp/tmp.JC94KtIlhb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kdjwuqdK1R ++ cat /tmp/tmp.JC94KtIlhb ++ rm /tmp/tmp.kdjwuqdK1R /tmp/tmp.JC94KtIlhb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 39 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LecN0hrNEs +++ mktemp ++ local LAST_ERR=/tmp/tmp.4Nm7dxUCCJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LecN0hrNEs ++ cat /tmp/tmp.4Nm7dxUCCJ ++ rm /tmp/tmp.LecN0hrNEs /tmp/tmp.4Nm7dxUCCJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 40 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KDQWiSyiED +++ mktemp ++ local LAST_ERR=/tmp/tmp.paYMfDyzlp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KDQWiSyiED ++ cat /tmp/tmp.paYMfDyzlp ++ rm /tmp/tmp.KDQWiSyiED /tmp/tmp.paYMfDyzlp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 41 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.umkp5zpwcl +++ mktemp ++ local LAST_ERR=/tmp/tmp.qAPFXDE52S ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.umkp5zpwcl ++ cat /tmp/tmp.qAPFXDE52S ++ rm /tmp/tmp.umkp5zpwcl /tmp/tmp.qAPFXDE52S ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 42 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N9ooL12wXe +++ mktemp ++ local LAST_ERR=/tmp/tmp.IixZ7eKSC6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.N9ooL12wXe ++ cat /tmp/tmp.IixZ7eKSC6 ++ rm /tmp/tmp.N9ooL12wXe /tmp/tmp.IixZ7eKSC6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 43 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5PLVbpF22H +++ mktemp ++ local LAST_ERR=/tmp/tmp.YjpsWotXWB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5PLVbpF22H ++ cat /tmp/tmp.YjpsWotXWB ++ rm /tmp/tmp.5PLVbpF22H /tmp/tmp.YjpsWotXWB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 44 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8KV0xIyOzj +++ mktemp ++ local LAST_ERR=/tmp/tmp.by97uOYfHP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8KV0xIyOzj ++ cat /tmp/tmp.by97uOYfHP ++ rm /tmp/tmp.8KV0xIyOzj /tmp/tmp.by97uOYfHP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 45 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.37hjyRVquo +++ mktemp ++ local LAST_ERR=/tmp/tmp.pDBdR6MXGp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.37hjyRVquo ++ cat /tmp/tmp.pDBdR6MXGp ++ rm /tmp/tmp.37hjyRVquo /tmp/tmp.pDBdR6MXGp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 46 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ly2VdTUhjQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.XfYL92IhC0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ly2VdTUhjQ ++ cat /tmp/tmp.XfYL92IhC0 ++ rm /tmp/tmp.Ly2VdTUhjQ /tmp/tmp.XfYL92IhC0 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wpszqks6In +++ mktemp ++ local LAST_ERR=/tmp/tmp.pRPhXM26KZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Wpszqks6In ++ cat /tmp/tmp.pRPhXM26KZ ++ rm /tmp/tmp.Wpszqks6In /tmp/tmp.pRPhXM26KZ ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.LdWMJFSdua ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.7UdYAXea0q +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.LdWMJFSdua +++++ cat /tmp/tmp.7UdYAXea0q +++++ rm /tmp/tmp.LdWMJFSdua /tmp/tmp.7UdYAXea0q +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ECU1a1nsRR +++ mktemp ++ local LAST_ERR=/tmp/tmp.5lnLWP205g ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ECU1a1nsRR ++ cat /tmp/tmp.5lnLWP205g ++ rm /tmp/tmp.ECU1a1nsRR /tmp/tmp.5lnLWP205g ++ return 0 + [[ 3 == \3 ]] + echo + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + check_generation 1 haproxy some-name + local generation=1 + local container=haproxy + local cluster=some-name + local current_generation ++ kubectl_bin get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.M1WAeDgPDp +++ mktemp ++ local LAST_ERR=/tmp/tmp.ddGpuHfBcL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.M1WAeDgPDp ++ cat /tmp/tmp.ddGpuHfBcL ++ rm /tmp/tmp.M1WAeDgPDp /tmp/tmp.ddGpuHfBcL ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.2dQv7rhbRd ++ mktemp + local LAST_ERR=/tmp/tmp.OX37llGbxh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2dQv7rhbRd secret/my-cluster-secrets patched + cat /tmp/tmp.OX37llGbxh + rm /tmp/tmp.2dQv7rhbRd /tmp/tmp.OX37llGbxh + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qVNbIQcjph +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z7xxYTvB85 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qVNbIQcjph ++ cat /tmp/tmp.Z7xxYTvB85 ++ rm /tmp/tmp.qVNbIQcjph /tmp/tmp.Z7xxYTvB85 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ccwNtek3Uy +++ mktemp ++ local LAST_ERR=/tmp/tmp.DtsaLeg79M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ccwNtek3Uy ++ cat /tmp/tmp.DtsaLeg79M ++ rm /tmp/tmp.ccwNtek3Uy /tmp/tmp.DtsaLeg79M ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.RJ02oTnD40 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.6ZWt9TUnM4 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.RJ02oTnD40 +++++ cat /tmp/tmp.6ZWt9TUnM4 +++++ rm /tmp/tmp.RJ02oTnD40 /tmp/tmp.6ZWt9TUnM4 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eO2vOyXu1e +++ mktemp ++ local LAST_ERR=/tmp/tmp.5tdcHEY6iD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eO2vOyXu1e ++ cat /tmp/tmp.5tdcHEY6iD ++ rm /tmp/tmp.eO2vOyXu1e /tmp/tmp.5tdcHEY6iD ++ return 0 + [[ 3 == \3 ]] + echo + sleep 15 + compare_mysql_cmd select-3 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-3-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8ldlXD8b9O +++ mktemp ++ local LAST_ERR=/tmp/tmp.LsdbZGF66Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8ldlXD8b9O ++ cat /tmp/tmp.LsdbZGF66Z ++ rm /tmp/tmp.8ldlXD8b9O /tmp/tmp.LsdbZGF66Z ++ return 0 + client_pod=pxc-client-56fd5498cd-qf49c + wait_pod pxc-client-56fd5498cd-qf49c + local pod=pxc-client-56fd5498cd-qf49c + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-qf49c ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-qf49c condition met waiting for pod/pxc-client-56fd5498cd-qf49c to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.QDkOK88Jf9/select-3.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.QDkOK88Jf9/select-3.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2434/e2e-tests/users-scheduler/compare/select-3.sql /tmp/tmp.QDkOK88Jf9/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + check_generation 1 haproxy some-name + local generation=1 + local container=haproxy + local cluster=some-name + local current_generation ++ kubectl_bin get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.j9v5SGuj4k +++ mktemp ++ local LAST_ERR=/tmp/tmp.zldOZ1tKxm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.j9v5SGuj4k ++ cat /tmp/tmp.zldOZ1tKxm ++ rm /tmp/tmp.j9v5SGuj4k /tmp/tmp.zldOZ1tKxm ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + destroy users-scheduler-5913 + local namespace=users-scheduler-5913 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v level=info + grep -v 'the object has been modified' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + grep -v 'get backup status: Job.batch' ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ + tee /tmp/tmp.QDkOK88Jf9/operator.log + sort -u +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ head -1 ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' +++ mktemp ++ local LAST_OUT=/tmp/tmp.me4zrxqejE +++ mktemp ++ local LAST_ERR=/tmp/tmp.sTrsiaawaB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.me4zrxqejE ++ cat /tmp/tmp.sTrsiaawaB ++ rm /tmp/tmp.me4zrxqejE /tmp/tmp.sTrsiaawaB ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-665ff7485-p85m7 ++ mktemp + local LAST_OUT=/tmp/tmp.3HZw2Pf21O ++ mktemp + local LAST_ERR=/tmp/tmp.t2PrIX9QQ6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-665ff7485-p85m7 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3HZw2Pf21O + cat /tmp/tmp.t2PrIX9QQ6 + rm /tmp/tmp.3HZw2Pf21O /tmp/tmp.t2PrIX9QQ6 + return 0 2026-04-09T05:19:25.816Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "", "enabled": ""} 2026-04-09T05:19:25.816Z INFO setup Manager starting up {"gitCommit": "3b65d7fe6a0a5db845238d34252d90db1f886e1d", "gitBranch": "PR-2434-3b65d7fe", "buildTime": "2026-04-09T01:11:10Z", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} 2026-04-09T05:19:25.816Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.32.13-gke.1205000"} 2026-04-09T05:19:25.819Z INFO setup Registering Components. 2026-04-09T05:19:26.502Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-04-09T05:19:26.503Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-04-09T05:19:26.503Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-04-09T05:19:26.503Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-04-09T05:19:26.503Z INFO controller-runtime.metrics Starting metrics server 2026-04-09T05:19:26.503Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-04-09T05:19:26.503Z INFO controller-runtime.webhook Starting webhook server 2026-04-09T05:19:26.503Z INFO setup Starting the Cmd. 2026-04-09T05:19:26.503Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-04-09T05:19:26.704Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-09T05:19:26.776Z DEBUG events percona-xtradb-cluster-operator-665ff7485-p85m7_c0194ca2-ee9e-4a45-a4e1-7f3e1d22578d became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"ebbed85d-2ff3-4970-bbaf-f7b1fac9c9d6","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1775711966739679009"}, "reason": "LeaderElection"} 2026-04-09T05:19:26.776Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-04-09T05:19:26.776Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-09T05:19:26.777Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-04-09T05:19:26.777Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-04-09T05:19:26.777Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-04-09T05:19:26.977Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-04-09T05:19:26.977Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-04-09T05:19:26.979Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-04-09T05:19:26.979Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-04-09T05:19:26.979Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-04-09T05:19:26.979Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-04-09T05:19:34.759Z ERROR Update status {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-2774"}, "namespace": "users-scheduler-2774", "name": "some-name", "reconcileID": "7440eb53-6302-4751-9061-b8ed0f7e9a32", "error": "PerconaXtraDBCluster.pxc.percona.com \"some-name\" not found"} 2026-04-09T05:20:03.554Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6a2d7efb-c0c7-49e5-990d-481b7fb6d216", "version": "1.20.0"} 2026-04-09T05:20:03.699Z INFO User secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6a2d7efb-c0c7-49e5-990d-481b7fb6d216", "secrets": "my-cluster-secrets"} 2026-04-09T05:20:03.916Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6a2d7efb-c0c7-49e5-990d-481b7fb6d216", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-09T05:20:03.938Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6a2d7efb-c0c7-49e5-990d-481b7fb6d216", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-09T05:20:04.501Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6a2d7efb-c0c7-49e5-990d-481b7fb6d216", "error": "failed to deploy proxysql: updatePod for proxysql: reconcile config: reconcile autotune config: create or update configmap: configmaps \"auto-some-name-pxc\" already exists", "errorVerbose": "configmaps \"auto-some-name-pxc\" already exists\ncreate or update configmap\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileAutotuneConfigMap\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:92\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:25\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:48\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:542\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile autotune config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:27\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:48\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:542\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:50\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:542\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nupdatePod for proxysql\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:543\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nfailed to deploy proxysql\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:561\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:20:04.617Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "915d5152-d48e-4e5e-8587-682cc2405e46", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-09T05:20:04.651Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "915d5152-d48e-4e5e-8587-682cc2405e46", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-09T05:20:04.705Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "915d5152-d48e-4e5e-8587-682cc2405e46", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-09T05:20:04.733Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "915d5152-d48e-4e5e-8587-682cc2405e46", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-09T05:20:04.781Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "915d5152-d48e-4e5e-8587-682cc2405e46", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-09T05:20:04.902Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "915d5152-d48e-4e5e-8587-682cc2405e46", "object": "some-name-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-09T05:20:05.606Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "106a8ee9-bceb-4bb4-b35c-0a19cadc9c7d", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-09T05:20:05.634Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "106a8ee9-bceb-4bb4-b35c-0a19cadc9c7d", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-09T05:21:22.186Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5", "user": "operator"} 2026-04-09T05:21:22.222Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5", "user": "monitor"} 2026-04-09T05:21:22.267Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5"} 2026-04-09T05:21:22.300Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5"} 2026-04-09T05:21:22.338Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5", "user": "xtrabackup"} 2026-04-09T05:21:22.381Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5"} 2026-04-09T05:21:22.418Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5", "user": "replication"} 2026-04-09T05:21:22.429Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "41f263a8-88dc-4bc8-92d1-69a58761bdd5", "err": "get primary pxc pod: not found"} 2026-04-09T05:21:27.536Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0833d37c-ef2a-4826-9f00-83862ec2ce7e", "err": "get primary pxc pod: not found"} 2026-04-09T05:21:32.651Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "01add60d-62cf-4218-a079-852338d5a8da", "err": "get primary pxc pod: not found"} 2026-04-09T05:24:03.939Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "cfb6e303-29bb-4bf8-b819-49999aae55cc", "user": "root"} 2026-04-09T05:24:04.002Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "cfb6e303-29bb-4bf8-b819-49999aae55cc", "new version": "8.4.7-7.1"} 2026-04-09T05:24:06.461Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "cfb6e303-29bb-4bf8-b819-49999aae55cc"} 2026-04-09T05:24:13.121Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9e9b756e-353f-4ce6-ab6a-1c595fc0119d"} 2026-04-09T05:24:18.267Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f1bc4091-4dfe-4186-b962-9cf48d67ebbf"} 2026-04-09T05:24:23.632Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "8a15fef3-2c23-40cd-9653-e2f2d4a6c88d"} 2026-04-09T05:24:28.835Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a5944027-e1bc-45c9-804c-d0fcad63995f"} 2026-04-09T05:24:33.949Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bf518cd0-71d5-4de8-8205-68c1655c1489"} 2026-04-09T05:24:39.420Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0593aa83-27fa-433d-9147-91c4a85e44dd"} 2026-04-09T05:24:44.438Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "46088f85-eb54-4dfd-9710-2bb4270c069b"} 2026-04-09T05:24:49.961Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f5e6bbdb-2654-4fb8-b3d0-528bb47ef727"} 2026-04-09T05:24:55.064Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "c58bbe57-1ecc-4261-83c2-65aa7d292818"} 2026-04-09T05:25:00.258Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "8e074fcb-5044-4690-bd60-29f134c8fc59"} 2026-04-09T05:25:05.693Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "972fd4d7-19d9-4c48-80cd-0c2d90c10e32"} 2026-04-09T05:25:11.071Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "95704d04-b780-4aab-956b-5b48ddfd8c19"} 2026-04-09T05:25:16.463Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "caa4c5bc-4087-42a2-accb-1524dcdcbdcf"} 2026-04-09T05:25:21.552Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e76424a0-68ab-43c3-881d-c11b005c7f04"} 2026-04-09T05:25:22.255Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a762f416-f9bb-4a0a-af32-13ac33393b92", "user": "root"} 2026-04-09T05:25:22.281Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a762f416-f9bb-4a0a-af32-13ac33393b92", "user": "root"} 2026-04-09T05:25:22.314Z INFO MySQL init secret created {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a762f416-f9bb-4a0a-af32-13ac33393b92", "secret": "some-name-mysql-init", "user": "root"} 2026-04-09T05:25:25.276Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a762f416-f9bb-4a0a-af32-13ac33393b92"} 2026-04-09T05:25:25.303Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a762f416-f9bb-4a0a-af32-13ac33393b92", "user": "root"} 2026-04-09T05:25:25.321Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a762f416-f9bb-4a0a-af32-13ac33393b92", "user": "root"} 2026-04-09T05:25:27.541Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a762f416-f9bb-4a0a-af32-13ac33393b92"} 2026-04-09T05:25:33.142Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "84217a03-51da-4251-9f21-ded26284906a"} 2026-04-09T05:25:38.333Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "dd9a6942-08e5-4f0d-aa9a-efdda477040e"} 2026-04-09T05:25:43.556Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "2f8c1922-4276-4ebb-92c7-5861ebebd4e6"} 2026-04-09T05:25:48.967Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "ff46d04b-86dc-493c-942c-a67c6cb733ec"} 2026-04-09T05:25:53.966Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f2b97502-61a0-4e4e-9610-fd54242c17f8"} 2026-04-09T05:25:59.313Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "73e3c0e4-cf0c-43b1-af1b-7f22b328528a"} 2026-04-09T05:26:04.947Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "088751a0-9d4d-4569-857e-9257c4c3def0"} 2026-04-09T05:26:05.275Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0f829a4a-ab52-407b-86e6-4c72eeae36a4", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:26:05.407Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0f829a4a-ab52-407b-86e6-4c72eeae36a4", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:26:08.884Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0f829a4a-ab52-407b-86e6-4c72eeae36a4", "error": "exec syncusers: failed to execute command in pod: pod some-name-proxysql-2 does not have a host assigned / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: pod some-name-proxysql-2 does not have a host assigned / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:26:29.781Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "16c44cfe-d106-4853-be28-c4a717bf9869", "user": "proxyadmin"} 2026-04-09T05:26:29.781Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "16c44cfe-d106-4853-be28-c4a717bf9869", "user": "proxyadmin"} 2026-04-09T05:26:29.845Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "16c44cfe-d106-4853-be28-c4a717bf9869", "user": "proxyadmin"} 2026-04-09T05:26:29.867Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "16c44cfe-d106-4853-be28-c4a717bf9869", "user": "proxyadmin"} 2026-04-09T05:26:29.867Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "16c44cfe-d106-4853-be28-c4a717bf9869", "last-applied-secret": "196c45042ebec515c6e135292b280dbd476987bd36153a167566947e3bbffc5e"} 2026-04-09T05:26:29.871Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "16c44cfe-d106-4853-be28-c4a717bf9869", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:26:33.051Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a3d53edd-d7b1-494d-a4dd-3e1b5e37a62f", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.50.41' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-1.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:26:47.610Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "00b13f67-bab2-4a5f-88c7-e39294ca43f5", "err": "get primary pxc pod: not found"} 2026-04-09T05:26:58.058Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "b94f287d-4667-4bf6-9944-58a58abfa87c", "err": "get primary pxc pod: not found"} 2026-04-09T05:27:17.676Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "3c8bf1dd-d75e-4df4-94b4-c4d2ae4f7ea1"} 2026-04-09T05:27:23.171Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "5181f39d-3ca3-484c-babb-12d9e7c1c09f"} 2026-04-09T05:27:28.412Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "89f7838d-83ad-4af9-a530-8edd0ea3f551"} 2026-04-09T05:27:33.881Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "713e4c90-e504-48e9-b12f-53d333c1349a"} 2026-04-09T05:27:35.119Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "d6ff7804-143f-41c2-9782-0ffa7bcf3026", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:27:35.300Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "d6ff7804-143f-41c2-9782-0ffa7bcf3026", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:27:37.606Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4d8a6c60-c338-467c-8050-bcb484d068ac", "user": "xtrabackup"} 2026-04-09T05:27:37.622Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4d8a6c60-c338-467c-8050-bcb484d068ac", "user": "xtrabackup"} 2026-04-09T05:27:37.639Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4d8a6c60-c338-467c-8050-bcb484d068ac", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-09T05:27:37.666Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4d8a6c60-c338-467c-8050-bcb484d068ac", "user": "xtrabackup"} 2026-04-09T05:27:37.677Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4d8a6c60-c338-467c-8050-bcb484d068ac", "user": "xtrabackup"} 2026-04-09T05:27:37.680Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4d8a6c60-c338-467c-8050-bcb484d068ac", "last-applied-secret": "e6c0342cfb43dda8e895000a36309ba9a1ad4281791af9b1b3d4686b503b265f"} 2026-04-09T05:27:37.683Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4d8a6c60-c338-467c-8050-bcb484d068ac", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:27:38.253Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "d6ff7804-143f-41c2-9782-0ffa7bcf3026"} 2026-04-09T05:28:42.587Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "946ca851-16e7-4d1a-9191-99142e0ce043", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: invalid connection"} 2026-04-09T05:29:35.061Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4950e84f-a2d1-4ed3-a258-7e16080fa4bb", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-5913 on 34.118.224.10:53: no such host"} 2026-04-09T05:29:40.380Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "407299bb-ac94-4bcd-a46c-d45eb66cfbc2", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:29:45.499Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "ebf0327e-586c-43ca-9a81-6eacf6b7c2f2", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:29:50.631Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "380fb588-b784-4e12-8211-2ec6f9d30869", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:29:55.756Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "1cb2d62a-e37d-47f7-aeca-00af6abaeba4", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:30:00.881Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4500d702-7259-4eeb-a7ec-6d1a74c84f7b", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:30:06.005Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "653eae3c-1166-451f-a4aa-50b68ff11f72", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:30:11.139Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "770a5c23-562d-46bb-8737-eb36ae8c2392", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:30:20.504Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f6adb39e-f169-49e0-b2c9-0d86ed54fd41"} 2026-04-09T05:30:24.845Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6230c1c1-ac71-4a35-9fdf-fa4e09e326ea"} 2026-04-09T05:30:30.450Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0915bcf9-4fb4-43e5-805e-61d7b6ac33b9"} 2026-04-09T05:30:35.188Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e8517478-52cf-4d4b-8ae0-681742478cbf"} 2026-04-09T05:30:37.025Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "user": "monitor"} 2026-04-09T05:30:37.039Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "user": "monitor"} 2026-04-09T05:30:37.065Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-09T05:30:37.088Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "user": "monitor"} 2026-04-09T05:30:37.115Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "user": "monitor"} 2026-04-09T05:30:37.403Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "last-applied-secret": "0aa5bf849ecc31345fb989e2a508b57fdce277f4d12db5bc20222e7a951194d0"} 2026-04-09T05:30:37.407Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:30:40.422Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f43a57c2-64c6-443b-a449-66c0f0f5998b", "error": "syncusers: ERROR (line:1612) : Could not find any nodes belonging to the cluster with writer hostgroup:11\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "syncusers: ERROR (line:1612) : Could not find any nodes belonging to the cluster with writer hostgroup:11\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:979\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:31:13.399Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "744e4d95-fa8c-4cae-a12c-1d65b7a2d7fa", "user": "monitor"} 2026-04-09T05:31:16.361Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "744e4d95-fa8c-4cae-a12c-1d65b7a2d7fa"} 2026-04-09T05:31:18.917Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "dddaa2dc-3829-4120-ba5b-61d3345d5346", "user": "monitor"} 2026-04-09T05:31:21.805Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "dddaa2dc-3829-4120-ba5b-61d3345d5346"} 2026-04-09T05:31:24.563Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e133a6d0-b366-4d4f-8382-b97330a4e212", "user": "monitor"} 2026-04-09T05:31:25.440Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e133a6d0-b366-4d4f-8382-b97330a4e212", "user": "monitor"} 2026-04-09T05:31:25.447Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e133a6d0-b366-4d4f-8382-b97330a4e212", "last-applied-secret": "0aa5bf849ecc31345fb989e2a508b57fdce277f4d12db5bc20222e7a951194d0"} 2026-04-09T05:31:27.810Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e133a6d0-b366-4d4f-8382-b97330a4e212"} 2026-04-09T05:31:33.095Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "c60a60e7-1c05-4835-91f2-ce5ae8f6ed98"} 2026-04-09T05:31:38.403Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e674b9f6-04a8-4e79-bcbc-22acf28cd798"} 2026-04-09T05:31:43.782Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "ff4653ea-8feb-406f-848d-28579c2405b7"} 2026-04-09T05:31:48.909Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6d8519c1-3bd3-4d23-b563-5e69cd3a5f2b"} 2026-04-09T05:31:54.494Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "b3d2d95f-1074-46a9-8955-74798e76b640"} 2026-04-09T05:31:55.671Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "user": "operator"} 2026-04-09T05:31:55.686Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "user": "operator"} 2026-04-09T05:31:55.708Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-09T05:31:55.729Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "user": "operator"} 2026-04-09T05:31:55.740Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "user": "operator"} 2026-04-09T05:31:55.752Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "last-applied-secret": "01c3b290ca7d32d4c6ada173b13143e006a93a1f1c0aa2f33e5f827025ed4897"} 2026-04-09T05:31:55.756Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:32:00.835Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6cd2358a-a765-4f9d-a6c7-2057edcbe7fa", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:32:04.687Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "88861d6c-eec9-45cb-b0a2-ce1761ee379f", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'10.139.48.77' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:32:18.665Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "13ae85a7-6b9e-469c-a257-f06642ffd3ca"} 2026-04-09T05:32:38.443Z INFO Created user secrets {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "secrets": "my-cluster-secrets-2"} 2026-04-09T05:32:38.449Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "root"} 2026-04-09T05:32:38.476Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "root"} 2026-04-09T05:32:38.494Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "secret": "some-name-mysql-init", "user": "root"} 2026-04-09T05:32:39.658Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4870f022-055a-418f-a324-e741a597fee2"} 2026-04-09T05:32:41.072Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a"} 2026-04-09T05:32:41.098Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "root"} 2026-04-09T05:32:41.116Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "root"} 2026-04-09T05:32:41.119Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "operator"} 2026-04-09T05:32:41.132Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "operator"} 2026-04-09T05:32:41.155Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-09T05:32:41.182Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "operator"} 2026-04-09T05:32:41.195Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "operator"} 2026-04-09T05:32:41.200Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "monitor"} 2026-04-09T05:32:41.215Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "monitor"} 2026-04-09T05:32:41.247Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-09T05:32:41.269Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "monitor"} 2026-04-09T05:32:41.334Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "monitor"} 2026-04-09T05:32:41.618Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "xtrabackup"} 2026-04-09T05:32:41.632Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "xtrabackup"} 2026-04-09T05:32:41.652Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-09T05:32:41.670Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "xtrabackup"} 2026-04-09T05:32:41.679Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "xtrabackup"} 2026-04-09T05:32:41.683Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "replication"} 2026-04-09T05:32:41.695Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "replication"} 2026-04-09T05:32:41.713Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "secret": "some-name-mysql-init", "user": "replication"} 2026-04-09T05:32:41.737Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "replication"} 2026-04-09T05:32:41.748Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "replication"} 2026-04-09T05:32:41.748Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "proxyadmin"} 2026-04-09T05:32:41.770Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "proxyadmin"} 2026-04-09T05:32:41.790Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "user": "proxyadmin"} 2026-04-09T05:32:41.790Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "last-applied-secret": "d839a6a57e717585d0ac3182a2331eccd6fb4c214f9548c6a4c7827116642cf4"} 2026-04-09T05:32:41.790Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "last-applied-secret": "d839a6a57e717585d0ac3182a2331eccd6fb4c214f9548c6a4c7827116642cf4"} 2026-04-09T05:32:41.795Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:32:41.854Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:32:45.869Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "345a4a4d-325d-45db-aa7d-c1d3c9be545a", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.78' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:33:41.316Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "236686f3-8ce9-401e-8235-4766f39a6df4", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.users-scheduler-5913 on 34.118.224.10:53: no such host"} 2026-04-09T05:33:52.104Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0afdf5af-95dd-4877-b177-45e520ebb17a", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp 10.139.49.99:33062: connect: connection refused"} 2026-04-09T05:34:50.718Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "c5510944-0658-44d3-a603-e0379522ab7c", "err": "failed to connect to pod some-name-pxc-0: invalid connection"} 2026-04-09T05:34:55.883Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "236feeac-0025-40d2-8226-36219c97e557", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:01.011Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "7caf0035-6977-4719-9e0d-f50de5a24a76", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:06.141Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "59630eb6-bba9-448f-abeb-d9c0f49f06dd", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:11.247Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "86f99c70-bc2f-4151-a9dc-65926d6e3fde", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:16.361Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "896f5aa7-e6cc-42fc-9989-f66f9529957e", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:21.473Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a93e5de6-9e77-4ce5-a950-c3aa45981be3", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:26.594Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "c453b5c8-2fb2-40a8-987b-1124dc546f0d", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:31.704Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0cd53ba1-78cf-484e-a6eb-b69453879309", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:36.833Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "afb2070e-6ab5-4fa9-bf6c-bd30e266d78d", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-5913.svc.cluster.local"} 2026-04-09T05:35:42.778Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "5a120c9f-a3cb-41fc-a6f7-448d240d3af7", "user": "monitor"} 2026-04-09T05:35:43.671Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "5a120c9f-a3cb-41fc-a6f7-448d240d3af7", "user": "monitor"} 2026-04-09T05:35:43.678Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "5a120c9f-a3cb-41fc-a6f7-448d240d3af7", "last-applied-secret": "d839a6a57e717585d0ac3182a2331eccd6fb4c214f9548c6a4c7827116642cf4"} 2026-04-09T05:35:45.649Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bac642e3-76b3-4f72-8cd8-b308baecebe6", "user": "operator"} 2026-04-09T05:35:45.664Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bac642e3-76b3-4f72-8cd8-b308baecebe6", "user": "operator"} 2026-04-09T05:35:45.684Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bac642e3-76b3-4f72-8cd8-b308baecebe6", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-09T05:35:45.709Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bac642e3-76b3-4f72-8cd8-b308baecebe6", "user": "operator"} 2026-04-09T05:35:45.721Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bac642e3-76b3-4f72-8cd8-b308baecebe6", "user": "operator"} 2026-04-09T05:35:45.735Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bac642e3-76b3-4f72-8cd8-b308baecebe6", "last-applied-secret": "dd7f40efd79b77a8143801aade35ec27dcb325f3d9cb684e10ab97231ea5e0f8"} 2026-04-09T05:35:45.742Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "bac642e3-76b3-4f72-8cd8-b308baecebe6", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:35:47.008Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "5a120c9f-a3cb-41fc-a6f7-448d240d3af7", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:35:55.047Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "5e62297c-8509-40db-8865-61ed01b0d935", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:36:16.546Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0e2bcdc1-a090-4630-abbc-d622065e5968"} 2026-04-09T05:36:22.113Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4b835902-3a6a-499a-8860-12aac1f4a66d"} 2026-04-09T05:36:27.453Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "07ed28b4-d9ff-4450-9640-e4cdec3b3f3f"} 2026-04-09T05:36:32.743Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "eb6c522a-1400-4097-8655-edb4f240cc5d"} 2026-04-09T05:36:37.746Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "ff190c4e-14c7-4310-aeb2-5b4884ce790b"} 2026-04-09T05:36:43.551Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9d0ba340-eaa0-45e1-bbdf-1c960652d0e6"} 2026-04-09T05:36:48.627Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "a35f2672-b51c-4e87-919a-f579a4e10196"} 2026-04-09T05:36:53.711Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "716efd61-7fd3-4339-9afd-f5ff8c7bb393"} 2026-04-09T05:36:59.238Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e6370b56-58dd-4e2c-828d-57c9dae70ab8"} 2026-04-09T05:37:04.439Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "ba4db5fa-0d92-4c35-bec2-12989efad3c0"} 2026-04-09T05:37:09.244Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "b3f313a3-a77d-4ba3-8c3c-a7bbc0e143fc"} 2026-04-09T05:37:15.118Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "83ecd284-f14c-4ec9-81bc-6b2fcce826ce"} 2026-04-09T05:37:20.189Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "e80c4c7f-7efa-48e2-b20f-46d134b128f9"} 2026-04-09T05:37:25.411Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "40c38ba4-75d9-4dfc-8f4a-9316e173b38c"} 2026-04-09T05:37:30.711Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "fc54ea7f-c577-4736-8fcb-c3a585fc6505"} 2026-04-09T05:37:35.907Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "fc0c60db-3041-401c-8bcf-58bd17b024d7"} 2026-04-09T05:37:41.236Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "1840dcbc-75f2-43cf-aa95-31afa0b295ed"} 2026-04-09T05:37:46.820Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "88f7e435-393c-4951-a69e-27d8ec00f063"} 2026-04-09T05:37:51.522Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "89f2200f-90fa-4ca5-8fcc-2ccd5d9632fb"} 2026-04-09T05:37:56.862Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "f19d905c-aca7-4a45-a8c3-4ac1d42c13a2"} 2026-04-09T05:38:02.126Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "56aab46a-b910-4318-9f65-7476ca7c6c5a"} 2026-04-09T05:38:06.664Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "root"} 2026-04-09T05:38:06.690Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "root"} 2026-04-09T05:38:06.717Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "secret": "some-name-mysql-init", "user": "root"} 2026-04-09T05:38:07.940Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "5c0494e8-bd25-4154-baee-a29410aaf148"} 2026-04-09T05:38:09.257Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0"} 2026-04-09T05:38:09.284Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "root"} 2026-04-09T05:38:09.301Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "root"} 2026-04-09T05:38:09.309Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "monitor"} 2026-04-09T05:38:09.322Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "monitor"} 2026-04-09T05:38:09.347Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-09T05:38:09.368Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "monitor"} 2026-04-09T05:38:09.399Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "monitor"} 2026-04-09T05:38:09.681Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "xtrabackup"} 2026-04-09T05:38:09.696Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "xtrabackup"} 2026-04-09T05:38:09.724Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-09T05:38:09.748Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "xtrabackup"} 2026-04-09T05:38:09.759Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "xtrabackup"} 2026-04-09T05:38:09.762Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "proxyadmin"} 2026-04-09T05:38:09.782Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "proxyadmin"} 2026-04-09T05:38:09.806Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "user": "proxyadmin"} 2026-04-09T05:38:09.806Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "last-applied-secret": "cef30302baad9c1db96dcf9f5c695fc2779c578e0e0293622dd11e8316dec3cf"} 2026-04-09T05:38:09.806Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "last-applied-secret": "cef30302baad9c1db96dcf9f5c695fc2779c578e0e0293622dd11e8316dec3cf"} 2026-04-09T05:38:09.808Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:38:09.887Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:38:13.422Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "659d5e25-9a21-4db7-82ab-bc7a62cacfe0", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.139.48.81' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-5913.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-09T05:38:18.948Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "ff1a50a8-17d7-40e5-b816-7411e93994e4", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp: lookup some-name-pxc-2.some-name-pxc.users-scheduler-5913 on 34.118.224.10:53: no such host"} 2026-04-09T05:39:07.867Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "4a39f2c0-04b8-4202-afb6-33b25628d5c0", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.users-scheduler-5913 on 34.118.224.10:53: no such host"} 2026-04-09T05:39:10.456Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9f95f96e-7ee9-44d3-9a0f-d18e34207d89", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-09T05:39:10.504Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9f95f96e-7ee9-44d3-9a0f-d18e34207d89", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-09T05:39:10.571Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9f95f96e-7ee9-44d3-9a0f-d18e34207d89", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-09T05:39:10.639Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9f95f96e-7ee9-44d3-9a0f-d18e34207d89", "object": "some-name-haproxy-replicas", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-09T05:39:19.211Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9f95f96e-7ee9-44d3-9a0f-d18e34207d89", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 34.118.233.185:3306: connect: connection refused"} 2026-04-09T05:39:19.357Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "c4f04156-4842-4277-b198-4899acfd0830", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-09T05:39:22.377Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "c4f04156-4842-4277-b198-4899acfd0830", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 34.118.233.185:3306: connect: connection refused"} 2026-04-09T05:39:30.522Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "9d785961-e46d-408e-a795-a4072753d2f4", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 34.118.233.185:3306: connect: connection refused"} 2026-04-09T05:40:56.316Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "root"} 2026-04-09T05:40:56.342Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "root"} 2026-04-09T05:40:56.367Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "secret": "some-name-mysql-init", "user": "root"} 2026-04-09T05:40:56.787Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "root"} 2026-04-09T05:40:56.805Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "root"} 2026-04-09T05:40:56.809Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "operator"} 2026-04-09T05:40:56.822Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "operator"} 2026-04-09T05:40:56.838Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-09T05:40:56.871Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "operator"} 2026-04-09T05:40:56.883Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "operator"} 2026-04-09T05:40:56.888Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "monitor"} 2026-04-09T05:40:56.901Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "monitor"} 2026-04-09T05:40:56.920Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-09T05:40:56.953Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "monitor"} 2026-04-09T05:40:57.311Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "xtrabackup"} 2026-04-09T05:40:57.324Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "xtrabackup"} 2026-04-09T05:40:57.345Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-09T05:40:57.365Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "xtrabackup"} 2026-04-09T05:40:57.375Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "xtrabackup"} 2026-04-09T05:40:57.380Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "replication"} 2026-04-09T05:40:57.396Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "replication"} 2026-04-09T05:40:57.418Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "secret": "some-name-mysql-init", "user": "replication"} 2026-04-09T05:40:57.441Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "replication"} 2026-04-09T05:40:57.453Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "user": "replication"} 2026-04-09T05:40:57.453Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "last-applied-secret": "01c3b290ca7d32d4c6ada173b13143e006a93a1f1c0aa2f33e5f827025ed4897"} 2026-04-09T05:40:57.455Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "6e2c3618-e7eb-466b-9606-0757b885bc53", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-09T05:42:01.813Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "35045fbc-6783-46ea-bd67-fd21b4c701f4", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.users-scheduler-5913 on 34.118.224.10:53: no such host"} 2026-04-09T05:42:07.840Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "15875655-7536-4601-980e-36f3a742fdc5", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: invalid connection"} 2026-04-09T05:43:01.177Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "434c4257-6a3e-4fe7-ac51-82fdde4c0d41", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-5913 on 34.118.224.10:53: no such host"} 2026-04-09T05:43:43.139Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "89df184f-5ec3-478b-b70c-41e9c00bed37", "user": "monitor"} 2026-04-09T05:43:44.056Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "89df184f-5ec3-478b-b70c-41e9c00bed37", "user": "monitor"} 2026-04-09T05:43:49.526Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "54b1da34-66bc-468f-bca6-e595f0d813e0", "user": "monitor"} 2026-04-09T05:43:49.540Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "54b1da34-66bc-468f-bca6-e595f0d813e0", "user": "monitor"} 2026-04-09T05:43:49.560Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "54b1da34-66bc-468f-bca6-e595f0d813e0", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-09T05:43:49.578Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "54b1da34-66bc-468f-bca6-e595f0d813e0", "user": "monitor"} 2026-04-09T05:43:55.002Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "39c581f7-d936-4b52-b7e7-14770c867ccd", "user": "monitor"} 2026-04-09T05:44:00.566Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "3e34b5fa-efde-4c74-b491-74e09bfc85f6", "user": "monitor"} 2026-04-09T05:44:05.934Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "384157bc-5bb4-4ed6-ae78-f6ff6344c4e4", "user": "monitor"} 2026-04-09T05:44:11.465Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "1010eafe-3ba7-4d6f-84a4-60f0905ed064", "user": "monitor"} 2026-04-09T05:44:16.965Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "0d80b15a-0d70-496a-9b8e-786c14f50ef5", "user": "monitor"} 2026-04-09T05:44:22.322Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "d4e6f836-a7a5-472e-8920-b38f5fb197a7", "user": "monitor"} 2026-04-09T05:44:27.906Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-5913"}, "namespace": "users-scheduler-5913", "name": "some-name", "reconcileID": "2286cce7-b8c4-437c-a914-7ced40467796", "user": "monitor"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile.func1 github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:259 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:484 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:834 [mysql] 2026/04/09 05:28:42 packets.go:58 read tcp 10.139.50.38:49542->10.139.49.96:33062: read: connection reset by peer [mysql] 2026/04/09 05:34:50 packets.go:58 read tcp 10.139.50.38:44932->10.139.50.43:33062: read: connection reset by peer [mysql] 2026/04/09 05:42:07 packets.go:58 read tcp 10.139.50.38:60156->10.139.49.102:33062: read: connection reset by peer sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 -  }, -  { -  }, +  }, -  "01c3b290ca7d32d4c6ada173b13143e006a93a1f1c0aa2f33e5f827025ed4897", +  "01c3b290ca7d32d4c6ada173b13143e006a93a1f1c0aa2f33e5f827025ed4897", +  "0aa5bf849ecc31345fb989e2a508b57fdce277f4d12db5bc20222e7a951194d0", -  "196c45042ebec515c6e135292b280dbd476987bd36153a167566947e3bbffc5e", +  "1c3b290ca7d32d4c6ada173b13143e006a93a1f1c0aa2f33e5f827025ed4897", -  "839a6a57e717585d0ac3182a2331eccd6fb4c214f9548c6a4c7827116642cf4", -  "aa5bf849ecc31345fb989e2a508b57fdce277f4d12db5bc20222e7a951194d0", -  Annotations: map[string]string{ +  Annotations: map[string]string{ -  APIVersion: "apps/v1", -  APIVersion: "apps/v1", +  AvailableReplicas: 0, -  AvailableReplicas: 2, -  AvailableReplicas: 3, -  "cef30302baad9c1db96dcf9f5c695fc2779c578e0e0293622dd11e8316dec3cf", +  "cef30302baad9c1db96dcf9f5c695fc2779c578e0e0293622dd11e8316dec3cf", -  CollisionCount: &0, +  CollisionCount: nil, +  CreationTimestamp: v1.Time{}, -  CreationTimestamp: v1.Time{Time: s"2026-04-09 05:20:04 +0000 UTC"}, +  CurrentReplicas: 0, -  CurrentReplicas: 2, -  CurrentReplicas: 3, +  CurrentRevision: "", -  CurrentRevision: "some-name-proxysql-558775d89c", -  CurrentRevision: "some-name-proxysql-65c97857fc", -  CurrentRevision: "some-name-proxysql-679bfbfd77", -  CurrentRevision: "some-name-proxysql-6c8474df4c", -  CurrentRevision: "some-name-proxysql-7fd64fd77f", -  CurrentRevision: "some-name-proxysql-86bdc56d58", -  CurrentRevision: "some-name-pxc-64744b7479", -  CurrentRevision: "some-name-pxc-67bc79f9fc", -  CurrentRevision: "some-name-pxc-7c6bd5999f", -  CurrentRevision: "some-name-pxc-8584fdb944", +  "d7f40efd79b77a8143801aade35ec27dcb325f3d9cb684e10ab97231ea5e0f8", -  "d839a6a57e717585d0ac3182a2331eccd6fb4c214f9548c6a4c7827116642cf4", +  "d839a6a57e717585d0ac3182a2331eccd6fb4c214f9548c6a4c7827116642cf4", -  "dd7f40efd79b77a8143801aade35ec27dcb325f3d9cb684e10ab97231ea5e0f8", -  DefaultMode: &420, -  DefaultMode: &420, +  DefaultMode: nil, +  DefaultMode: nil, +  DeprecatedServiceAccount: "", -  DeprecatedServiceAccount: "default", +  DNSPolicy: "", -  DNSPolicy: "ClusterFirst", -  "e6c0342cfb43dda8e895000a36309ba9a1ad4281791af9b1b3d4686b503b265f", -  FieldsType: "FieldsV1", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., +  Generation: 0, -  Generation: 1, -  Generation: 2, -  Generation: 3, -  Generation: 4, -  Generation: 5, -  Generation: 6, -  Generation: 7, -  Generation: 8, +  "last-applied-secret": "196c45042ebec515c6e135292b280dbd476987bd36153a167566947e3bbffc5e", +  "last-applied-secret": "e6c0342cfb43dda8e895000a36309ba9a1ad4281791af9b1b3d4686b503b265f", +  ManagedFields: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  Manager: "kube-controller-manager", -  Manager: "percona-xtradb-cluster-operator", +  ObservedGeneration: 0, -  ObservedGeneration: 1, -  ObservedGeneration: 2, -  ObservedGeneration: 3, -  ObservedGeneration: 4, -  ObservedGeneration: 5, -  ObservedGeneration: 6, -  ObservedGeneration: 7, -  ObservedGeneration: 8, -  Operation: "Update", -  Operation: "Update", -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMDFjM2IyOTBjYTdkMzJkNGM2YWRhMTczYjEzMTQzZTAwNmE5M2ExZjFjMGFhMmYzM2U1ZjgyNzAyNWVkNDg5NyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMDFjM2IyOTBjYTdkMzJkNGM2YWRhMTczYjEzMTQzZTAwNmE5M2ExZjFjMGFhMmYzM2U1ZjgyNzAyNWVkNDg5NyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMGFhNWJmODQ5ZWNjMzEzNDVmYjk4OWUyYTUwOGI1N2ZkY2UyNzdmNGQxMmRiNWJjMjAyMjJlN2E5NTExOTRkMCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMGFhNWJmODQ5ZWNjMzEzNDVmYjk4OWUyYTUwOGI1N2ZkY2UyNzdmNGQxMmRiNWJjMjAyMjJlN2E5NTExOTRkMCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMTk2YzQ1MDQyZWJlYzUxNWM2ZTEzNTI5MmIyODBkYmQ0NzY5ODdiZDM2MTUzYTE2NzU2Njk0N2UzYmJmZmM1ZSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiY2VmMzAzMDJiYWFkOWMxZGI5NmRjZjlmNWM2OTVmYzI3NzljNTc4ZTBlMDI5MzYyMmRkMTFlODMxNmRlYzNjZiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZDgzOWE2YTU3ZTcxNzU4NWQwYWMzMTgyYTIzMzFlY2NkNmZiNGMyMTRmOTU0OGM2YTRjNzgyNzExNjY0MmNmNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZDgzOWE2YTU3ZTcxNzU4NWQwYWMzMTgyYTIzMzFlY2NkNmZiNGMyMTRmOTU0OGM2YTRjNzgyNzExNjY0MmNmNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZGQ3ZjQwZWZkNzliNzdhODE0MzgwMWFhZGUzNWVjMjdkY2IzMjVmM2Q5Y2I2ODRlMTBhYjk3MjMxZWE1ZTBmOCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZGQ3ZjQwZWZkNzliNzdhODE0MzgwMWFhZGUzNWVjMjdkY2IzMjVmM2Q5Y2I2ODRlMTBhYjk3MjMxZWE1ZTBmOCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMDFjM2IyOTBjYTdkMzJkNGM2YWRhMTczYjEzMTQzZTAwNmE5M2ExZjFjMGFhMmYzM2U1ZjgyNzAyNWVkNDg5NyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiY2VmMzAzMDJiYWFkOWMxZGI5NmRjZjlmNWM2OTVmYzI3NzljNTc4ZTBlMDI5MzYyMmRkMTFlODMxNmRlYzNjZiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiY2VmMzAzMDJiYWFkOWMxZGI5NmRjZjlmNWM2OTVmYzI3NzljNTc4ZTBlMDI5MzYyMmRkMTFlODMxNmRlYzNjZiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZDgzOWE2YTU3ZTcxNzU4NWQwYWMzMTgyYTIzMzFlY2NkNmZiNGMyMTRmOTU0OGM2YTRjNzgyNzExNjY0MmNmNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZDgzOWE2YTU3ZTcxNzU4NWQwYWMzMTgyYTIzMzFlY2NkNmZiNGMyMTRmOTU0OGM2YTRjNzgyNzExNjY0MmNmNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZTZjMDM0MmNmYjQzZGRhOGU4OTUwMDBhMzYzMDliYTlhMWFkNDI4MTc5MWFmOWIxYjNkNDY4NmI1MDNiMjY1ZiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZTZjMDM0MmNmYjQzZGRhOGU4OTUwMDBhMzYzMDliYTlhMWFkNDI4MTc5MWFmOWIxYjNkNDY4NmI1MDNiMjY1ZiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJwZXJjb25hLmNvbS9jb25maWd1cmF0aW9uLWhhc2giOiJkNDFkOGNkOThmMDBiMjA0ZTk4MDA5OThlY2Y4NDI3ZSIsInBlcmNvbmEuY29tL3NzbC1oYXNoIjoiY2ZkNmE1MjM5ODI2ODE3M2I1MWQ3Y2RiMTMzMWMwOWEiLCJwZXJjb25hLmNv"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMTk2YzQ1MDQyZWJlYzUxNWM2ZTEzNTI5MmIyODBkYmQ0NzY5ODdiZDM2MTUzYTE2NzU2Njk0N2UzYmJmZmM1ZSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJwZXJjb25hLmNvbS9jb25maWd1cmF0aW9uLWhhc2giOiJkNDFkOGNkOThmMDBiMjA0ZTk4MDA5OThlY2Y4NDI3ZSIsInBlcmNvbmEuY29tL3NzbC1oYXNoIjoiY2ZkNmE1MjM5ODI2ODE3M2I1MWQ3Y2RiMTMzMWMw"..., +  PeriodSeconds: 0, -  PeriodSeconds: 10, +  PersistentVolumeClaimRetentionPolicy: nil, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  Phase: "", -  Phase: "Pending", +  PodManagementPolicy: "", -  PodManagementPolicy: "OrderedReady", +  Protocol: "", -  Protocol: "TCP", +  ReadyReplicas: 0, -  ReadyReplicas: 2, -  ReadyReplicas: 3, +  Replicas: 0, -  Replicas: 2, -  Replicas: &2, +  Replicas: &2, -  Replicas: 3, -  Replicas: &3, +  Replicas: &3, +  ResourceVersion: "", -  ResourceVersion: "1775712042256079004", -  ResourceVersion: "1775712237941039014", -  ResourceVersion: "1775712386786975004", -  ResourceVersion: "1775712428293791004", -  ResourceVersion: "1775712456273695004", -  ResourceVersion: "1775712616107807014", -  ResourceVersion: "1775712664648591004", -  ResourceVersion: "1775712734773199004", -  ResourceVersion: "1775712789894991004", -  ResourceVersion: "1775712937610783014", -  ResourceVersion: "1775712973165007004", -  ResourceVersion: "1775713254547919014", +  RestartPolicy: "", -  RestartPolicy: "Always", -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil, +  SchedulerName: "", -  SchedulerName: "default-scheduler", -  Subresource: "status", +  TerminationMessagePath: "", -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePolicy: "", -  TerminationMessagePolicy: "File", -  Time: s"2026-04-09 05:20:04 +0000 UTC", -  Time: s"2026-04-09 05:20:42 +0000 UTC", -  Time: s"2026-04-09 05:23:57 +0000 UTC", -  Time: s"2026-04-09 05:26:05 +0000 UTC", -  Time: s"2026-04-09 05:26:26 +0000 UTC", -  Time: s"2026-04-09 05:26:29 +0000 UTC", -  Time: s"2026-04-09 05:27:08 +0000 UTC", -  Time: s"2026-04-09 05:27:35 +0000 UTC", -  Time: s"2026-04-09 05:27:36 +0000 UTC", -  Time: s"2026-04-09 05:27:37 +0000 UTC", -  Time: s"2026-04-09 05:30:16 +0000 UTC", -  Time: s"2026-04-09 05:30:37 +0000 UTC", -  Time: s"2026-04-09 05:31:04 +0000 UTC", -  Time: s"2026-04-09 05:31:55 +0000 UTC", -  Time: s"2026-04-09 05:32:14 +0000 UTC", -  Time: s"2026-04-09 05:32:41 +0000 UTC", -  Time: s"2026-04-09 05:33:09 +0000 UTC", -  Time: s"2026-04-09 05:35:37 +0000 UTC", -  Time: s"2026-04-09 05:35:45 +0000 UTC", -  Time: s"2026-04-09 05:36:13 +0000 UTC", -  Time: s"2026-04-09 05:38:09 +0000 UTC", -  Time: s"2026-04-09 05:40:54 +0000 UTC", -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, +  UID: "", -  UID: "4c6e89a6-eb64-4560-99f7-4d5537de5e96", -  UID: "77efeab9-bf1c-4bbc-9fe8-60c1569a5b41", +  UpdatedReplicas: 0, -  UpdatedReplicas: 2, -  UpdatedReplicas: 3, +  UpdateRevision: "", -  UpdateRevision: "some-name-proxysql-558775d89c", -  UpdateRevision: "some-name-proxysql-65c97857fc", -  UpdateRevision: "some-name-proxysql-679bfbfd77", -  UpdateRevision: "some-name-proxysql-6c8474df4c", -  UpdateRevision: "some-name-proxysql-7fd64fd77f", -  UpdateRevision: "some-name-proxysql-86bdc56d58", -  UpdateRevision: "some-name-pxc-64744b7479", -  UpdateRevision: "some-name-pxc-67bc79f9fc", -  UpdateRevision: "some-name-pxc-7c6bd5999f", -  UpdateRevision: "some-name-pxc-8584fdb944", -  VolumeMode: &"Filesystem", +  VolumeMode: nil,   }    },    },    {    },    },    {    },    }, ""),    },    {    },    "0",    ... // 16 identical fields    ... // 16 identical fields    ... // 22 identical fields    ... // 2 identical fields    ... // 3 identical fields    ... // 3 identical fields    ... // 3 identical fields    ... // 4 identical fields    ... // 5 identical fields    ... // 6 identical fields    ... // 7 identical fields    ... // 8 identical fields    ... // 9 identical fields    ... // 9 identical fields    AccessModes: nil,    ActiveDeadlineSeconds: nil,    Affinity: nil,    Annotations: map[string]string{    Args: {"mysqld"},    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    AutomountServiceAccountToken: nil,    AWSElasticBlockStore: nil,    AzureFile: nil,    Capacity: nil,    Conditions: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    ContainerPort: 3306,    ContainerPort: 33060,    ContainerPort: 33062,    ContainerPort: 4444,    ContainerPort: 4567,    ContainerPort: 4568,    ContainerPort: 6032,    ContainerPort: 6070,    Containers: []v1.Container{    "d",    DataSource: nil,    DataSourceRef: nil,    DeletionGracePeriodSeconds: nil,    DeletionGracePeriodSeconds: nil,    DeletionTimestamp: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-env-vars-proxysql"}, Optional: &true}}},    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "some-name-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "proxyadmin"}}}, ...},    Env: {{Name: "PXC_SERVICE", Value: "some-name-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "xtrabackup"}}}, ...},    EphemeralContainers: nil,    FailureThreshold: 3,    FC: nil,    Finalizers: nil,    Finalizers: nil,    GitRepo: nil,    HostAliases: nil,    HostIP: "",    HostPort: 0,    ImagePullPolicy: "Always",    InitContainers: []v1.Container{    InitialDelaySeconds: 300,    ISCSI: nil,    Items: nil,    Items: nil,    "kubectl.kubernetes.io/default-container": "proxysql",    "kubectl.kubernetes.io/default-container": "pxc",    Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Labels: nil,    "last-applied-secret": strings.Join({    Lifecycle: nil,    LivenessProbe: &v1.Probe{    LocalObjectReference: {Name: "auto-some-name-pxc"},    LocalObjectReference: {Name: "some-name-pxc"},    ManagedFields: nil,    MinReadySeconds: 0,    Name: "auto-config",    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    Name: "config",    Name: "ist",    Name: "mysql",    Name: "mysql-admin",    Name: "mysql-init-file",    Name: "mysql-users-secret-file",    Name: "mysqlx",    Name: "proxyadm",    Namespace: "users-scheduler-5913",    Name: "ssl",    Name: "ssl-internal",    Name: "sst",    Name: "stats",    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    Name: "vault-keyring-secret",    Name: "write-set",    NFS: nil,    NodeName: "",    NodeSelector: nil,    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "last-applied-secret": "196c45042ebec515c6e135292b280dbd476987bd36153a167566947e3bbffc5e", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", ...}},    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc"}},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    ObjectMeta: v1.ObjectMeta{    ObjectMeta: v1.ObjectMeta{    Optional: &false,    Optional: &true,    Optional: &true,    Ordinals: nil,    OS: nil,    Overhead: nil,    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "some-name", UID: "ed1e4eba-a6b4-4c90-97e6-9cac4d9da026", ...}},    OwnerReferences: nil,    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a",    "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a",    "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc",    "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc",    Ports: []v1.ContainerPort{    PreemptionPolicy: nil,    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    Quobyte: nil,    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    Replicas: &2,    Replicas: &3,    SecretName: "internal-some-name",    SecretName: "some-name-mysql-init",    SecretName: "some-name-ssl",    SecretName: "some-name-ssl-internal",    SecretName: "some-name-vault",    Secret: &v1.SecretVolumeSource{    SecurityContext: nil,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    SelfLink: "",    ServiceAccountName: "default",    ServiceName: "some-name-proxysql-unready",    ServiceName: "some-name-pxc",    SetHostnameAsFQDN: nil,    Spec: v1.PersistentVolumeClaimSpec{    Spec: v1.PodSpec{    Spec: v1.StatefulSetSpec{    StartupProbe: nil,    Status: v1.PersistentVolumeClaimStatus{    Status: v1.StatefulSetStatus{    StorageClassName: nil,    Subdomain: "",    SuccessThreshold: 1,    Template: v1.PodTemplateSpec{    TerminationGracePeriodSeconds: &30,    TerminationGracePeriodSeconds: &600,    TerminationGracePeriodSeconds: nil,    TimeoutSeconds: 5,    Tolerations: nil,    TypeMeta: {},    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}},   &v1.StatefulSet{    VolumeAttributesClassName: nil,    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    VolumeDevices: nil,    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeName: "",    VolumeSource: v1.VolumeSource{    Volumes: []v1.Volume{    VsphereVolume: nil,    WorkingDir: "", + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch pxc -n users-scheduler-5913 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.FpyAuYc6lE ++ mktemp + local LAST_ERR=/tmp/tmp.wFArz5FViT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FpyAuYc6lE perconaxtradbcluster.pxc.percona.com "some-name" deleted from users-scheduler-5913 namespace + cat /tmp/tmp.wFArz5FViT + rm /tmp/tmp.FpyAuYc6lE /tmp/tmp.wFArz5FViT + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.eUlAmHGqCR ++ mktemp + local LAST_ERR=/tmp/tmp.6uLiclRYNi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eUlAmHGqCR No resources found + cat /tmp/tmp.6uLiclRYNi + rm /tmp/tmp.eUlAmHGqCR /tmp/tmp.6uLiclRYNi + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.J9PD6CSGgi ++ mktemp + local LAST_ERR=/tmp/tmp.oKit7aC5lg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.J9PD6CSGgi No resources found + cat /tmp/tmp.oKit7aC5lg + rm /tmp/tmp.J9PD6CSGgi /tmp/tmp.oKit7aC5lg + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.v3FCPANOo8 ++ mktemp + local LAST_ERR=/tmp/tmp.gokzbaEvPi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.v3FCPANOo8 validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.gokzbaEvPi + rm /tmp/tmp.v3FCPANOo8 /tmp/tmp.gokzbaEvPi + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace users-scheduler-5913 + rm -rf /tmp/tmp.QDkOK88Jf9 + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.8WG4QXXTOZ + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- ++ mktemp + local LAST_OUT=/tmp/tmp.lnQe2HZpLI ++ mktemp + local LAST_ERR=/tmp/tmp.7htgsc9doa + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.j2hk79VTAB + local exit_status=0 + for i in '$(seq 0 2)' + set +e ++ seq 0 2 + kubectl delete --grace-period=0 --force=true namespace users-scheduler-5913 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator