Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/logs/users-scheduler-8-4.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + create_infra users-scheduler-10069 + local ns=users-scheduler-10069 + '[' -n pxc-operator ']' + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified + : + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.EDh05lxHbc ++ mktemp + local LAST_ERR=/tmp/tmp.XYjoYcqavk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EDh05lxHbc No resources found + cat /tmp/tmp.XYjoYcqavk + rm /tmp/tmp.EDh05lxHbc /tmp/tmp.XYjoYcqavk + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.VKkI3sFkXw ++ mktemp + local LAST_ERR=/tmp/tmp.7XoLms5Z6p + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VKkI3sFkXw No resources found + cat /tmp/tmp.7XoLms5Z6p + rm /tmp/tmp.VKkI3sFkXw /tmp/tmp.7XoLms5Z6p + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.7rpHwfUZcu ++ mktemp + local LAST_ERR=/tmp/tmp.4o3fGqyQrs + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.7rpHwfUZcu No resources found + cat /tmp/tmp.4o3fGqyQrs + rm /tmp/tmp.7rpHwfUZcu /tmp/tmp.4o3fGqyQrs + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ awk '{print $1}' ++pod/percona-xtradb-cluster-operator-5b5dc68764-jdg9h condition met waiting for pod/percona-xtradb-cluster-operator-5b5dc68764-jdg9h to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces users-scheduler-31309 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "users-scheduler-31309" not found waiting for namespace/users-scheduler-31309 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "users-scheduler-31309" not found ----------------------------------------------------------------------------------- create namespace users-scheduler-31309 ----------------------------------------------------------------------------------- namespace/users-scheduler-31309 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2427-d4f7121d-1-cluster3" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created ----------------------------------------------------------------------------------- create PXC cluster with 1-password secret ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/some-name created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- error: timed out waiting for the condition on pods/some-name-pxc-0 waiting for pod/some-name-pxc-0 to become Ready.Ok pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ERROR 1045 (28000): ProxySQL Error: Access denied for user 'root'@'10.148.74.62' (using password: YES) command terminated with exit code 1 pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ERROR 1062 (23000) at line 1: Duplicate entry '100500' for key 'myApp.PRIMARY' command terminated with exit code 1 pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok Unable to use a TTY - input is not a terminal or the right kind of file ----------------------------------------------------------------------------------- test missing passwords were created and present in internal secrets ----------------------------------------------------------------------------------- Checking root Running compare for root pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok Checking xtrabackup Running compare for xtrabackup pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok Checking monitor Running compare for monitor pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok Checking proxyadmin Running compare for proxyadmin Checking operator Running compare for operator pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok Checking replication Running compare for replication pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- secret/my-cluster-secrets patched (no change) ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/some-name to be ready.......................... pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/some-name patched ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/some-name to be ready..... secret/my-cluster-secrets patched (no change) ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/some-name to be ready ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/some-name patched secret/my-cluster-secrets patched (no change) ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/some-name to be ready ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- secret/my-cluster-secrets patched (no change) pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok waiting for password update pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok {"additional_password": "$A$005$r?\\u0018\\u0001S\\u000bY&_B[,ev\\u0001WD@\\f\\u001dw8wbkgIBsfIyRk1SEJlw.rK3f9xyLeloAcusBFQao77"} waiting for old password to be discarded... (retry 10/600) waiting for old password to be discarded... (retry 20/600) waiting for old password to be discarded... (retry 30/600) waiting for old password to be discarded... (retry 40/600) local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4VbYjly8XN secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.TvhKIDDCTj + rm /tmp/tmp.4VbYjly8XN /tmp/tmp.TvhKIDDCTj + return 0 + desc 'create PXC cluster with 1-password secret' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster with 1-password secret ----------------------------------------------------------------------------------- + newpass=test-password ++ echo -n test-password ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZA== + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/some-name.yml '' '' /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/secrets_one_pass.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/secrets_one_pass.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/secrets_one_pass.yml ++ mktemp + local LAST_OUT=/tmp/tmp.7IlMi5MtKJ ++ mktemp + local LAST_ERR=/tmp/tmp.fWUxXtJxSN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/secrets_one_pass.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.7IlMi5MtKJ secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.fWUxXtJxSN + rm /tmp/tmp.7IlMi5MtKJ /tmp/tmp.fWUxXtJxSN + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/client.yml + local pvc_name= + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-scheduler-10069~ + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.CroMvlOTFu ++ mktemp + local LAST_ERR=/tmp/tmp.YD0tscbiu2 + local exit_status=0 ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2427-d4f7121d#' + for i in '$(seq 0 2)' + set +e + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + kubectl apply -f - + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CroMvlOTFu deployment.apps/pxc-client created + cat /tmp/tmp.YD0tscbiu2 + rm /tmp/tmp.CroMvlOTFu /tmp/tmp.YD0tscbiu2 + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/some-name.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/some-name.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/some-name.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/some-name.yml + local pvc_name= + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2427-d4f7121d#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.users-scheduler-10069~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' ++ mktemp + local LAST_OUT=/tmp/tmp.JePEqvN3ti ++ mktemp + local LAST_ERR=/tmp/tmp.Yic69UhlI8 + local exit_status=0 + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JePEqvN3ti perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.Yic69UhlI8 + rm /tmp/tmp.JePEqvN3ti /tmp/tmp.Yic69UhlI8 + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1D4v1QtRla ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gg2XnwdEC2 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.1D4v1QtRla +++ cat /tmp/tmp.gg2XnwdEC2 +++ rm /tmp/tmp.1D4v1QtRla /tmp/tmp.gg2XnwdEC2 +++ return 0 ++ [[ false == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.4TQx76Cziv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.WHYeCwWFS7 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.4TQx76Cziv +++ cat /tmp/tmp.WHYeCwWFS7 +++ rm /tmp/tmp.4TQx76Cziv /tmp/tmp.WHYeCwWFS7 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-10069 ++ mktemp + local LAST_OUT=/tmp/tmp.AxDxaaW87p ++ mktemp + local LAST_ERR=/tmp/tmp.TxMcbDC9cr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-10069 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-10069 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n users-scheduler-10069 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.AxDxaaW87p + cat /tmp/tmp.TxMcbDC9cr error: no matching resources found + rm /tmp/tmp.AxDxaaW87p /tmp/tmp.TxMcbDC9cr + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.kKQL478yp4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.dRoePimCul ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kKQL478yp4 ++ cat /tmp/tmp.dRoePimCul ++ rm /tmp/tmp.kKQL478yp4 /tmp/tmp.dRoePimCul ++ return 0 + local 'root_pass=N.a)#f%2p$q(q]gZ%' + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dF5tCi8K7h +++ mktemp ++ local LAST_ERR=/tmp/tmp.JALK8R7Uis ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dF5tCi8K7h ++ cat /tmp/tmp.JALK8R7Uis ++ rm /tmp/tmp.dF5tCi8K7h /tmp/tmp.JALK8R7Uis ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZrnB2Zo8g2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.aLDyRhXS7f ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZrnB2Zo8g2 ++ cat /tmp/tmp.aLDyRhXS7f ++ rm /tmp/tmp.ZrnB2Zo8g2 /tmp/tmp.aLDyRhXS7f ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZIQVXA6NHM +++ mktemp ++ local LAST_ERR=/tmp/tmp.W6PUD151yU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZIQVXA6NHM ++ cat /tmp/tmp.W6PUD151yU ++ rm /tmp/tmp.ZIQVXA6NHM /tmp/tmp.W6PUD151yU ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1.sql /tmp/tmp.uv48lQ0yvW/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7AjUurvsi4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.kzd6uHlMGh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7AjUurvsi4 ++ cat /tmp/tmp.kzd6uHlMGh ++ rm /tmp/tmp.7AjUurvsi4 /tmp/tmp.kzd6uHlMGh ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1.sql /tmp/tmp.uv48lQ0yvW/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''N.a)#f%2p$q(q]gZ%'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6HRDQm5Tff +++ mktemp ++ local LAST_ERR=/tmp/tmp.1TzGow7SyN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6HRDQm5Tff ++ cat /tmp/tmp.1TzGow7SyN ++ rm /tmp/tmp.6HRDQm5Tff /tmp/tmp.1TzGow7SyN ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-2fzfm + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-1.sql /tmp/tmp.uv48lQ0yvW/select-1.sql + is_keyring_plugin_in_use some-name + local cluster=some-name + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + kubectl exec some-name-pxc-0 -c pxc -- ls /var/lib/mysql/mysqld.my + return 1 ++ kubectl exec -it some-name-proxysql-0 -- sh -c 'proxysql --version 2>/dev/null' ++ awk '{print $3}' ++ cut -d. -f1 Unable to use a TTY - input is not a terminal or the right kind of file + PROXYSQL_VER=2 + tables_cmp_file=select-2-proxysql3 + [[ 2 == 2 ]] + tables_cmp_file=select-2 + desc 'test missing passwords were created and present in internal secrets' + set +o xtrace ----------------------------------------------------------------------------------- test missing passwords were created and present in internal secrets ----------------------------------------------------------------------------------- + empty_pwds=() + wrong_pwds=() + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking root' Checking root ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ft0p8tSpue +++ mktemp ++ local LAST_ERR=/tmp/tmp.7fMVE81LO4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ft0p8tSpue ++ cat /tmp/tmp.7fMVE81LO4 ++ rm /tmp/tmp.Ft0p8tSpue /tmp/tmp.7fMVE81LO4 ++ return 0 + secret_pass='N.a)#f%2p$q(q]gZ%' ++ getSecretData internal-some-name root ++ local secretName=internal-some-name ++ local dataKey=root ++ base64 --decode ++ kubectl_bin get secrets/internal-some-name '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KKkt3T5hyk +++ mktemp ++ local LAST_ERR=/tmp/tmp.V0NFQsi04E ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KKkt3T5hyk ++ cat /tmp/tmp.V0NFQsi04E ++ rm /tmp/tmp.KKkt3T5hyk /tmp/tmp.V0NFQsi04E ++ return 0 + int_secret_pass='N.a)#f%2p$q(q]gZ%' + [[ -z N.a)#f%2p$q(q]gZ% ]] + [[ N.a)#f%2p$q(q]gZ% != \N\.\a\)\#\f\%\2\p\$\q\(\q\]\g\Z\% ]] + [[ root != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ root ]] + [[ '' =~ root ]] + echo 'Running compare for root' Running compare for root + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''N.a)#f%2p$q(q]gZ%'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GNWFBbYd8V +++ mktemp ++ local LAST_ERR=/tmp/tmp.BLri0jzEoZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GNWFBbYd8V ++ cat /tmp/tmp.BLri0jzEoZ ++ rm /tmp/tmp.GNWFBbYd8V /tmp/tmp.BLri0jzEoZ ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking xtrabackup' Checking xtrabackup ++ getSecretData my-cluster-secrets xtrabackup ++ local secretName=my-cluster-secrets ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.R2OeuRoefo +++ mktemp ++ local LAST_ERR=/tmp/tmp.BaCmvj0Eqq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.R2OeuRoefo ++ cat /tmp/tmp.BaCmvj0Eqq ++ rm /tmp/tmp.R2OeuRoefo /tmp/tmp.BaCmvj0Eqq ++ return 0 + secret_pass='El=l2C' ++ getSecretData internal-some-name xtrabackup ++ local secretName=internal-some-name ++ local dataKey=xtrabackup ++ kubectl_bin get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.TrO4YCKieX +++ mktemp ++ local LAST_ERR=/tmp/tmp.NVsDlSPzlC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.xtrabackup}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TrO4YCKieX ++ cat /tmp/tmp.NVsDlSPzlC ++ rm /tmp/tmp.TrO4YCKieX /tmp/tmp.NVsDlSPzlC ++ return 0 + int_secret_pass='El=l2C' + [[ -z El=l2C ]] + [[ El=l2C != \E\<\m\L\5\C\]\M\E\&\5\[\D\>\l\=\l\2\C ]] + [[ xtrabackup != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ xtrabackup ]] + [[ '' =~ xtrabackup ]] + echo 'Running compare for xtrabackup' Running compare for xtrabackup + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\''El=l2C'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\''El=l2C'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uxtrabackup -p'\''El=l2C'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uxtrabackup -p'\''El=l2C'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.irW7AmvFC8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tkvCZYB5mG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.irW7AmvFC8 ++ cat /tmp/tmp.tkvCZYB5mG ++ rm /tmp/tmp.irW7AmvFC8 /tmp/tmp.tkvCZYB5mG ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-56fd5498cd-2fzfm ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking monitor' Checking monitor ++ getSecretData my-cluster-secrets monitor ++ local secretName=my-cluster-secrets ++ local dataKey=monitor ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.gBmCg7AzcZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.JRaw8KR2Pa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gBmCg7AzcZ ++ cat /tmp/tmp.JRaw8KR2Pa ++ rm /tmp/tmp.gBmCg7AzcZ /tmp/tmp.JRaw8KR2Pa ++ return 0 + secret_pass=monitor_password ++ getSecretData internal-some-name monitor ++ local secretName=internal-some-name ++ local dataKey=monitor ++ kubectl_bin get secrets/internal-some-name '--template={{.data.monitor}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.P3WJe8WDLw +++ mktemp ++ local LAST_ERR=/tmp/tmp.pujZT1oNB1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.monitor}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.P3WJe8WDLw ++ cat /tmp/tmp.pujZT1oNB1 ++ rm /tmp/tmp.P3WJe8WDLw /tmp/tmp.pujZT1oNB1 ++ return 0 + int_secret_pass=monitor_password + [[ -z monitor_password ]] + [[ monitor_password != \m\o\n\i\t\o\r\_\p\a\s\s\w\o\r\d ]] + [[ monitor != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ monitor ]] + [[ '' =~ monitor ]] + echo 'Running compare for monitor' Running compare for monitor + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''monitor_password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''monitor_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VAXgiF6LyJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.JbsYDNYqwG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VAXgiF6LyJ ++ cat /tmp/tmp.JbsYDNYqwG ++ rm /tmp/tmp.VAXgiF6LyJ /tmp/tmp.JbsYDNYqwG ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking proxyadmin' Checking proxyadmin ++ getSecretData my-cluster-secrets proxyadmin ++ local secretName=my-cluster-secrets ++ local dataKey=proxyadmin ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.kqYNNxti6q +++ mktemp ++ local LAST_ERR=/tmp/tmp.6oJJHakHyU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.proxyadmin}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kqYNNxti6q ++ cat /tmp/tmp.6oJJHakHyU ++ rm /tmp/tmp.kqYNNxti6q /tmp/tmp.6oJJHakHyU ++ return 0 + secret_pass='I%s!K{PEYS<_R}77m' ++ getSecretData internal-some-name proxyadmin ++ local secretName=internal-some-name ++ local dataKey=proxyadmin ++ kubectl_bin get secrets/internal-some-name '--template={{.data.proxyadmin}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ddjorRB1Qq +++ mktemp ++ local LAST_ERR=/tmp/tmp.4JPfoiYGVc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.proxyadmin}}' ++ base64 --decode ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ddjorRB1Qq ++ cat /tmp/tmp.4JPfoiYGVc ++ rm /tmp/tmp.ddjorRB1Qq /tmp/tmp.4JPfoiYGVc ++ return 0 + int_secret_pass='I%s!K{PEYS<_R}77m' + [[ -z I%s!K{PEYS<_R}77m ]] + [[ I%s!K{PEYS<_R}77m != \I\%\s\!\K\{\P\E\Y\S\<\_\R\}\7\7\m ]] + [[ proxyadmin != \p\r\o\x\y\a\d\m\i\n ]] + [[ proxyadmin == \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ proxyadmin ]] + [[ '' =~ proxyadmin ]] + echo 'Running compare for proxyadmin' Running compare for proxyadmin + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''I%s!K{PEYS<_R}77m'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''I%s!K{PEYS<_R}77m'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''I%s!K{PEYS<_R}77m'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''I%s!K{PEYS<_R}77m'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.uv48lQ0yvW/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.uv48lQ0yvW/select-2.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking operator' Checking operator ++ getSecretData my-cluster-secrets operator ++ local secretName=my-cluster-secrets ++ local dataKey=operator ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.61aIoC4d0K +++ mktemp ++ local LAST_ERR=/tmp/tmp.E4QSfpKgfN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.61aIoC4d0K ++ cat /tmp/tmp.E4QSfpKgfN ++ rm /tmp/tmp.61aIoC4d0K /tmp/tmp.E4QSfpKgfN ++ return 0 + secret_pass='vDR@g+c$w@rOs7Z1m!R' ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.465X03Jbbg +++ mktemp ++ local LAST_ERR=/tmp/tmp.n7D8WvFBJM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.465X03Jbbg ++ cat /tmp/tmp.n7D8WvFBJM ++ rm /tmp/tmp.465X03Jbbg /tmp/tmp.n7D8WvFBJM ++ return 0 + int_secret_pass='vDR@g+c$w@rOs7Z1m!R' + [[ -z vDR@g+c$w@rOs7Z1m!R ]] + [[ vDR@g+c$w@rOs7Z1m!R != \v\D\R\@\g\+\c\$\w\@\r\O\s\7\Z\1\m\!\R ]] + [[ operator != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ operator ]] + [[ '' =~ operator ]] + echo 'Running compare for operator' Running compare for operator + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''vDR@g+c$w@rOs7Z1m!R'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''vDR@g+c$w@rOs7Z1m!R'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''vDR@g+c$w@rOs7Z1m!R'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''vDR@g+c$w@rOs7Z1m!R'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.19DycfpEHa +++ mktemp ++ local LAST_ERR=/tmp/tmp.BJnVgMure2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.19DycfpEHa ++ cat /tmp/tmp.BJnVgMure2 ++ rm /tmp/tmp.19DycfpEHa /tmp/tmp.BJnVgMure2 ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-56fd5498cd-2fzfm ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + for user in root xtrabackup monitor proxyadmin operator replication + echo 'Checking replication' Checking replication ++ getSecretData my-cluster-secrets replication ++ local secretName=my-cluster-secrets ++ local dataKey=replication ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.replication}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.BalWgQi6T0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CbPugAqB3j ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.replication}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BalWgQi6T0 ++ cat /tmp/tmp.CbPugAqB3j ++ rm /tmp/tmp.BalWgQi6T0 /tmp/tmp.CbPugAqB3j ++ return 0 + secret_pass=',X-f=xy2{~,HzXhr' ++ getSecretData internal-some-name replication ++ local secretName=internal-some-name ++ local dataKey=replication ++ kubectl_bin get secrets/internal-some-name '--template={{.data.replication}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.ODpUNJovT9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.5mei4iHUjq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.replication}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ODpUNJovT9 ++ cat /tmp/tmp.5mei4iHUjq ++ rm /tmp/tmp.ODpUNJovT9 /tmp/tmp.5mei4iHUjq ++ return 0 + int_secret_pass=',X-f=xy2{~,HzXhr' + [[ -z ,X-f=xy2{~,HzXhr ]] + [[ ,X-f=xy2{~,HzXhr != \,\X\-\f\=\x\y\2\{\~\,\H\z\X\h\r ]] + [[ replication != \p\r\o\x\y\a\d\m\i\n ]] + [[ '' =~ replication ]] + [[ '' =~ replication ]] + echo 'Running compare for replication' Running compare for replication + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\'',X-f=xy2{~,HzXhr'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\'',X-f=xy2{~,HzXhr'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -ureplication -p'\'',X-f=xy2{~,HzXhr'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -ureplication -p'\'',X-f=xy2{~,HzXhr'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LYo1HnMnxX +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lcr0TgeHYL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LYo1HnMnxX ++ cat /tmp/tmp.Lcr0TgeHYL ++ rm /tmp/tmp.LYo1HnMnxX /tmp/tmp.Lcr0TgeHYL ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ -n '' ]] + desc 'test root' + set +o xtrace ----------------------------------------------------------------------------------- test root ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets root dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=root + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.hkSV8yjbSw ++ mktemp + local LAST_ERR=/tmp/tmp.9z602LTfxG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"root": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hkSV8yjbSw secret/my-cluster-secrets patched + cat /tmp/tmp.9z602LTfxG + rm /tmp/tmp.hkSV8yjbSw /tmp/tmp.9z602LTfxG + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tB2LgV9v1p +++ mktemp ++ local LAST_ERR=/tmp/tmp.nDMcjMsfbo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tB2LgV9v1p ++ cat /tmp/tmp.nDMcjMsfbo ++ rm /tmp/tmp.tB2LgV9v1p /tmp/tmp.nDMcjMsfbo ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6eXjoWDvCr +++ mktemp ++ local LAST_ERR=/tmp/tmp.raPf36FBEV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6eXjoWDvCr ++ cat /tmp/tmp.raPf36FBEV ++ rm /tmp/tmp.6eXjoWDvCr /tmp/tmp.raPf36FBEV ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.fk3YtcJIqG ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.E3pHBlQwTf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.fk3YtcJIqG +++++ cat /tmp/tmp.E3pHBlQwTf +++++ rm /tmp/tmp.fk3YtcJIqG /tmp/tmp.E3pHBlQwTf +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.r7Fi1yhLGp ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.KdZgQOoEJF +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.r7Fi1yhLGp +++++ cat /tmp/tmp.KdZgQOoEJF +++++ rm /tmp/tmp.r7Fi1yhLGp /tmp/tmp.KdZgQOoEJF +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GPauZqcUxs +++ mktemp ++ local LAST_ERR=/tmp/tmp.ORrJkAXS7c ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GPauZqcUxs ++ cat /tmp/tmp.ORrJkAXS7c ++ rm /tmp/tmp.GPauZqcUxs /tmp/tmp.ORrJkAXS7c ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uroot -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t7jJi82hGJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.s6C7LWrA2r ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.t7jJi82hGJ ++ cat /tmp/tmp.s6C7LWrA2r ++ rm /tmp/tmp.t7jJi82hGJ /tmp/tmp.s6C7LWrA2r ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + desc 'test proxyadmin' + set +o xtrace ----------------------------------------------------------------------------------- test proxyadmin ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.4lwsBlr6lb ++ mktemp + local LAST_ERR=/tmp/tmp.YzrCiLqVpN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":3}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4lwsBlr6lb perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.YzrCiLqVpN + rm /tmp/tmp.4lwsBlr6lb /tmp/tmp.YzrCiLqVpN + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gmTVFLrbrz +++ mktemp ++ local LAST_ERR=/tmp/tmp.Hayk09DwKH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gmTVFLrbrz ++ cat /tmp/tmp.Hayk09DwKH ++ rm /tmp/tmp.gmTVFLrbrz /tmp/tmp.Hayk09DwKH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wO1y3ExoMp +++ mktemp ++ local LAST_ERR=/tmp/tmp.3WiauQSHO9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wO1y3ExoMp ++ cat /tmp/tmp.3WiauQSHO9 ++ rm /tmp/tmp.wO1y3ExoMp /tmp/tmp.3WiauQSHO9 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aoLQpUgyWi +++ mktemp ++ local LAST_ERR=/tmp/tmp.vs0WwptPO7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aoLQpUgyWi ++ cat /tmp/tmp.vs0WwptPO7 ++ rm /tmp/tmp.aoLQpUgyWi /tmp/tmp.vs0WwptPO7 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.CAM0dL3RfB ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.pkoSig9dWJ +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.CAM0dL3RfB +++++ cat /tmp/tmp.pkoSig9dWJ +++++ rm /tmp/tmp.CAM0dL3RfB /tmp/tmp.pkoSig9dWJ +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.MY05Q8ZgZv ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.sRDCYKwuzF +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.MY05Q8ZgZv +++++ cat /tmp/tmp.sRDCYKwuzF +++++ rm /tmp/tmp.MY05Q8ZgZv /tmp/tmp.sRDCYKwuzF +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ny9FyBZvYW +++ mktemp ++ local LAST_ERR=/tmp/tmp.7ixD34ZyNf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ny9FyBZvYW ++ cat /tmp/tmp.7ixD34ZyNf ++ rm /tmp/tmp.Ny9FyBZvYW /tmp/tmp.7ixD34ZyNf ++ return 0 + [[ 3 == \3 ]] + echo + patch_secret my-cluster-secrets proxyadmin dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=proxyadmin + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.uc4yXbTYg1 ++ mktemp + local LAST_ERR=/tmp/tmp.WxPs8gsqxE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"proxyadmin": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uc4yXbTYg1 secret/my-cluster-secrets patched + cat /tmp/tmp.WxPs8gsqxE + rm /tmp/tmp.uc4yXbTYg1 /tmp/tmp.WxPs8gsqxE + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IHAqpJ6maG +++ mktemp ++ local LAST_ERR=/tmp/tmp.tghjQjzi4r ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IHAqpJ6maG ++ cat /tmp/tmp.tghjQjzi4r ++ rm /tmp/tmp.IHAqpJ6maG /tmp/tmp.tghjQjzi4r ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KSubXFjOvY +++ mktemp ++ local LAST_ERR=/tmp/tmp.vGMP2BztVw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KSubXFjOvY ++ cat /tmp/tmp.vGMP2BztVw ++ rm /tmp/tmp.KSubXFjOvY /tmp/tmp.vGMP2BztVw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Mub2Y0t3Yr +++ mktemp ++ local LAST_ERR=/tmp/tmp.xqg9uMC8yn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Mub2Y0t3Yr ++ cat /tmp/tmp.xqg9uMC8yn ++ rm /tmp/tmp.Mub2Y0t3Yr /tmp/tmp.xqg9uMC8yn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K98vCohD4A +++ mktemp ++ local LAST_ERR=/tmp/tmp.Bb1yVIAHPC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.K98vCohD4A ++ cat /tmp/tmp.Bb1yVIAHPC ++ rm /tmp/tmp.K98vCohD4A /tmp/tmp.Bb1yVIAHPC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EAw62AHqxQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jf8LBf3U4u ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EAw62AHqxQ ++ cat /tmp/tmp.Jf8LBf3U4u ++ rm /tmp/tmp.EAw62AHqxQ /tmp/tmp.Jf8LBf3U4u ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gftz8bQzn4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.NVfARvajBN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gftz8bQzn4 ++ cat /tmp/tmp.NVfARvajBN ++ rm /tmp/tmp.gftz8bQzn4 /tmp/tmp.NVfARvajBN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mscnAIdISA +++ mktemp ++ local LAST_ERR=/tmp/tmp.xkT5fxk18Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mscnAIdISA ++ cat /tmp/tmp.xkT5fxk18Z ++ rm /tmp/tmp.mscnAIdISA /tmp/tmp.xkT5fxk18Z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z9LIRBO4kV +++ mktemp ++ local LAST_ERR=/tmp/tmp.k7FOIBjuI3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Z9LIRBO4kV ++ cat /tmp/tmp.k7FOIBjuI3 ++ rm /tmp/tmp.Z9LIRBO4kV /tmp/tmp.k7FOIBjuI3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MOn8oDmgtP +++ mktemp ++ local LAST_ERR=/tmp/tmp.VXqYJAUp4n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MOn8oDmgtP ++ cat /tmp/tmp.VXqYJAUp4n ++ rm /tmp/tmp.MOn8oDmgtP /tmp/tmp.VXqYJAUp4n ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TjYKgDXCo5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.A5zBhlgS0Q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TjYKgDXCo5 ++ cat /tmp/tmp.A5zBhlgS0Q ++ rm /tmp/tmp.TjYKgDXCo5 /tmp/tmp.A5zBhlgS0Q ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.cW0aI9Q5tx ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.QfL23BMDee +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.cW0aI9Q5tx +++++ cat /tmp/tmp.QfL23BMDee +++++ rm /tmp/tmp.cW0aI9Q5tx /tmp/tmp.QfL23BMDee +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.NXFIf5qSp5 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.SYUUW9LWt8 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.NXFIf5qSp5 +++++ cat /tmp/tmp.SYUUW9LWt8 +++++ rm /tmp/tmp.NXFIf5qSp5 /tmp/tmp.SYUUW9LWt8 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bB6MT3oHTW +++ mktemp ++ local LAST_ERR=/tmp/tmp.8QJuCyJ21B ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bB6MT3oHTW ++ cat /tmp/tmp.8QJuCyJ21B ++ rm /tmp/tmp.bB6MT3oHTW /tmp/tmp.8QJuCyJ21B ++ return 0 + [[ 3 == \3 ]] + echo + sleep 15 + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-0 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-0 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.uv48lQ0yvW/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.uv48lQ0yvW/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.uv48lQ0yvW/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.uv48lQ0yvW/select-2.sql + compare_mysql_cmd_local select-2 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=select-2 + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2-84.sql ']' + run_mysql_local 'SHOW TABLES;' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SHOW TABLES;' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.uv48lQ0yvW/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-2.sql /tmp/tmp.uv48lQ0yvW/select-2.sql + compare_mysql_cmd_local proxysql-cfg 'SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 '' proxysql + local command_id=proxysql-cfg + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/proxysql-cfg.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/proxysql-cfg-84.sql ']' + run_mysql_local 'SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-2 proxysql + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''admin-cluster_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-2 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.uv48lQ0yvW/proxysql-cfg.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/proxysql-cfg.sql /tmp/tmp.uv48lQ0yvW/proxysql-cfg.sql + desc 'test xtrabackup' + set +o xtrace ----------------------------------------------------------------------------------- test xtrabackup ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.APkDvnwzdJ ++ mktemp + local LAST_ERR=/tmp/tmp.2AqdWvT3FJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge '-p={"spec":{"proxysql":{"size":2}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.APkDvnwzdJ perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.2AqdWvT3FJ + rm /tmp/tmp.APkDvnwzdJ /tmp/tmp.2AqdWvT3FJ + return 0 + patch_secret my-cluster-secrets xtrabackup dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=xtrabackup + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.vEyA3YS9fU ++ mktemp + local LAST_ERR=/tmp/tmp.W4De8qbpOn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vEyA3YS9fU secret/my-cluster-secrets patched + cat /tmp/tmp.W4De8qbpOn + rm /tmp/tmp.vEyA3YS9fU /tmp/tmp.W4De8qbpOn + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dSTEvxhQvU +++ mktemp ++ local LAST_ERR=/tmp/tmp.mCtQ4vyWHx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dSTEvxhQvU ++ cat /tmp/tmp.mCtQ4vyWHx ++ rm /tmp/tmp.dSTEvxhQvU /tmp/tmp.mCtQ4vyWHx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z1W6CLh22F +++ mktemp ++ local LAST_ERR=/tmp/tmp.DOppHSRE19 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z1W6CLh22F ++ cat /tmp/tmp.DOppHSRE19 ++ rm /tmp/tmp.z1W6CLh22F /tmp/tmp.DOppHSRE19 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mBvjzpuI37 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DHAQmtGdxf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mBvjzpuI37 ++ cat /tmp/tmp.DHAQmtGdxf ++ rm /tmp/tmp.mBvjzpuI37 /tmp/tmp.DHAQmtGdxf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BLUmpZkkTH +++ mktemp ++ local LAST_ERR=/tmp/tmp.Xt9nRsJE6a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BLUmpZkkTH ++ cat /tmp/tmp.Xt9nRsJE6a ++ rm /tmp/tmp.BLUmpZkkTH /tmp/tmp.Xt9nRsJE6a ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ClSlNJUbqX +++ mktemp ++ local LAST_ERR=/tmp/tmp.EhxddL9aNc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ClSlNJUbqX ++ cat /tmp/tmp.EhxddL9aNc ++ rm /tmp/tmp.ClSlNJUbqX /tmp/tmp.EhxddL9aNc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ds8tqYHyLp +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q7NyFKtogF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ds8tqYHyLp ++ cat /tmp/tmp.Q7NyFKtogF ++ rm /tmp/tmp.ds8tqYHyLp /tmp/tmp.Q7NyFKtogF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cPs4H6kM73 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FK0V32AlrQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cPs4H6kM73 ++ cat /tmp/tmp.FK0V32AlrQ ++ rm /tmp/tmp.cPs4H6kM73 /tmp/tmp.FK0V32AlrQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7Bk9Vg1m4s +++ mktemp ++ local LAST_ERR=/tmp/tmp.1SOkPrYlhK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7Bk9Vg1m4s ++ cat /tmp/tmp.1SOkPrYlhK ++ rm /tmp/tmp.7Bk9Vg1m4s /tmp/tmp.1SOkPrYlhK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nXYhgfuuWV +++ mktemp ++ local LAST_ERR=/tmp/tmp.LaFZawbN4K ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nXYhgfuuWV ++ cat /tmp/tmp.LaFZawbN4K ++ rm /tmp/tmp.nXYhgfuuWV /tmp/tmp.LaFZawbN4K ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UmTcgC0Crz +++ mktemp ++ local LAST_ERR=/tmp/tmp.jA4whSg0Bt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UmTcgC0Crz ++ cat /tmp/tmp.jA4whSg0Bt ++ rm /tmp/tmp.UmTcgC0Crz /tmp/tmp.jA4whSg0Bt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.075MOTBBs0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HGa29wLgEJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.075MOTBBs0 ++ cat /tmp/tmp.HGa29wLgEJ ++ rm /tmp/tmp.075MOTBBs0 /tmp/tmp.HGa29wLgEJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ynVeVdqe02 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gnX9gdG5TK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ynVeVdqe02 ++ cat /tmp/tmp.gnX9gdG5TK ++ rm /tmp/tmp.ynVeVdqe02 /tmp/tmp.gnX9gdG5TK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JNhSkrXmmP +++ mktemp ++ local LAST_ERR=/tmp/tmp.YlytkDZerM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JNhSkrXmmP ++ cat /tmp/tmp.YlytkDZerM ++ rm /tmp/tmp.JNhSkrXmmP /tmp/tmp.YlytkDZerM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XB8NWv6611 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZrzbNJlxEm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XB8NWv6611 ++ cat /tmp/tmp.ZrzbNJlxEm ++ rm /tmp/tmp.XB8NWv6611 /tmp/tmp.ZrzbNJlxEm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tRZF4iwGAL +++ mktemp ++ local LAST_ERR=/tmp/tmp.8Fr4Ij5Lrq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tRZF4iwGAL ++ cat /tmp/tmp.8Fr4Ij5Lrq ++ rm /tmp/tmp.tRZF4iwGAL /tmp/tmp.8Fr4Ij5Lrq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5n0m4IVLJr +++ mktemp ++ local LAST_ERR=/tmp/tmp.JeVtkA2IjX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5n0m4IVLJr ++ cat /tmp/tmp.JeVtkA2IjX ++ rm /tmp/tmp.5n0m4IVLJr /tmp/tmp.JeVtkA2IjX ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5mCE3GwkyF +++ mktemp ++ local LAST_ERR=/tmp/tmp.8yVedjR6bl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5mCE3GwkyF ++ cat /tmp/tmp.8yVedjR6bl ++ rm /tmp/tmp.5mCE3GwkyF /tmp/tmp.8yVedjR6bl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qZ5squw0vH +++ mktemp ++ local LAST_ERR=/tmp/tmp.DGnmQABDCr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qZ5squw0vH ++ cat /tmp/tmp.DGnmQABDCr ++ rm /tmp/tmp.qZ5squw0vH /tmp/tmp.DGnmQABDCr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oFoPmXcUhJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.7KF7PgEpIo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oFoPmXcUhJ ++ cat /tmp/tmp.7KF7PgEpIo ++ rm /tmp/tmp.oFoPmXcUhJ /tmp/tmp.7KF7PgEpIo ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pL6tDVpjwA +++ mktemp ++ local LAST_ERR=/tmp/tmp.ol3LK5vSdG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pL6tDVpjwA ++ cat /tmp/tmp.ol3LK5vSdG ++ rm /tmp/tmp.pL6tDVpjwA /tmp/tmp.ol3LK5vSdG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jLOG4WkSBc +++ mktemp ++ local LAST_ERR=/tmp/tmp.9F81RcJAvB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jLOG4WkSBc ++ cat /tmp/tmp.9F81RcJAvB ++ rm /tmp/tmp.jLOG4WkSBc /tmp/tmp.9F81RcJAvB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e6uALseGDY +++ mktemp ++ local LAST_ERR=/tmp/tmp.a343xwNwIy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.e6uALseGDY ++ cat /tmp/tmp.a343xwNwIy ++ rm /tmp/tmp.e6uALseGDY /tmp/tmp.a343xwNwIy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mV6RNiZOKR +++ mktemp ++ local LAST_ERR=/tmp/tmp.4Coad7UOQu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mV6RNiZOKR ++ cat /tmp/tmp.4Coad7UOQu ++ rm /tmp/tmp.mV6RNiZOKR /tmp/tmp.4Coad7UOQu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VWYgiyeE50 +++ mktemp ++ local LAST_ERR=/tmp/tmp.eJOvI90tLE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VWYgiyeE50 ++ cat /tmp/tmp.eJOvI90tLE ++ rm /tmp/tmp.VWYgiyeE50 /tmp/tmp.eJOvI90tLE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rk3vKbNnkZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.aP2QUUD9sN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rk3vKbNnkZ ++ cat /tmp/tmp.aP2QUUD9sN ++ rm /tmp/tmp.rk3vKbNnkZ /tmp/tmp.aP2QUUD9sN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.skxpQgJ8Uo +++ mktemp ++ local LAST_ERR=/tmp/tmp.lbqE8BNQfm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.skxpQgJ8Uo ++ cat /tmp/tmp.lbqE8BNQfm ++ rm /tmp/tmp.skxpQgJ8Uo /tmp/tmp.lbqE8BNQfm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S9Hpfn4Vgd +++ mktemp ++ local LAST_ERR=/tmp/tmp.doyzf73oAa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S9Hpfn4Vgd ++ cat /tmp/tmp.doyzf73oAa ++ rm /tmp/tmp.S9Hpfn4Vgd /tmp/tmp.doyzf73oAa ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 26 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oemLOQJTmd +++ mktemp ++ local LAST_ERR=/tmp/tmp.8bAFndmW6y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oemLOQJTmd ++ cat /tmp/tmp.8bAFndmW6y ++ rm /tmp/tmp.oemLOQJTmd /tmp/tmp.8bAFndmW6y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 27 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.btyLRX6tny +++ mktemp ++ local LAST_ERR=/tmp/tmp.kQVmtnyIMa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.btyLRX6tny ++ cat /tmp/tmp.kQVmtnyIMa ++ rm /tmp/tmp.btyLRX6tny /tmp/tmp.kQVmtnyIMa ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 28 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.akKiiNlzln +++ mktemp ++ local LAST_ERR=/tmp/tmp.rsXknXQiSr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.akKiiNlzln ++ cat /tmp/tmp.rsXknXQiSr ++ rm /tmp/tmp.akKiiNlzln /tmp/tmp.rsXknXQiSr ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ypXkBOmKqO +++ mktemp ++ local LAST_ERR=/tmp/tmp.iCgS7BhvIG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ypXkBOmKqO ++ cat /tmp/tmp.iCgS7BhvIG ++ rm /tmp/tmp.ypXkBOmKqO /tmp/tmp.iCgS7BhvIG ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.LvRIgslYso ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.7T3Ruro7qa +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.LvRIgslYso +++++ cat /tmp/tmp.7T3Ruro7qa +++++ rm /tmp/tmp.LvRIgslYso /tmp/tmp.7T3Ruro7qa +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.dFfujzm4od ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Snxw8x4eVb +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.dFfujzm4od +++++ cat /tmp/tmp.Snxw8x4eVb +++++ rm /tmp/tmp.dFfujzm4od /tmp/tmp.Snxw8x4eVb +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jraqaKZdRp +++ mktemp ++ local LAST_ERR=/tmp/tmp.dZq9P02KT3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jraqaKZdRp ++ cat /tmp/tmp.dZq9P02KT3 ++ rm /tmp/tmp.jraqaKZdRp /tmp/tmp.dZq9P02KT3 ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd_local select-3 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 '' pxc + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-3-84.sql ']' + run_mysql_local 'SHOW DATABASES;' '-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' some-name-pxc-0 pxc + local 'command=SHOW DATABASES;' + local 'uri=-h 127.0.0.1 -uxtrabackup -p'\''test-password'\''' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + '[' '!' -s /tmp/tmp.uv48lQ0yvW/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-3.sql /tmp/tmp.uv48lQ0yvW/select-3.sql + desc 'test monitor' + set +o xtrace ----------------------------------------------------------------------------------- test monitor ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.6M4tffNLP4 ++ mktemp + local LAST_ERR=/tmp/tmp.WjS33yBeAW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6M4tffNLP4 secret/my-cluster-secrets patched + cat /tmp/tmp.WjS33yBeAW + rm /tmp/tmp.6M4tffNLP4 /tmp/tmp.WjS33yBeAW + return 0 + wait_for_password_propagation my-cluster-secrets monitor some-name + local secret=my-cluster-secrets + local user=monitor + local cluster=some-name + local max_retry=600 + local root_pass + local retry ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.nCA060EJ9S +++ mktemp ++ local LAST_ERR=/tmp/tmp.nFoMzMm5V8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nCA060EJ9S ++ cat /tmp/tmp.nFoMzMm5V8 ++ rm /tmp/tmp.nCA060EJ9S /tmp/tmp.nFoMzMm5V8 ++ return 0 + root_pass=test-password + [[ -z test-password ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + retry=0 + is_password_updated monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + local 'command=SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' + grep additional_password + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eyOBd8pFsU +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ec49Op8D8h ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eyOBd8pFsU ++ cat /tmp/tmp.Ec49Op8D8h ++ rm /tmp/tmp.eyOBd8pFsU /tmp/tmp.Ec49Op8D8h ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} + sleep 5 + retry=0 + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=1 + [[ 1 -eq 0 ]] + [[ 1 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=2 + [[ 2 -eq 0 ]] + [[ 2 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=3 + [[ 3 -eq 0 ]] + [[ 3 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=4 + [[ 4 -eq 0 ]] + [[ 4 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=5 + [[ 5 -eq 0 ]] + [[ 5 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=6 + [[ 6 -eq 0 ]] + [[ 6 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=7 + [[ 7 -eq 0 ]] + [[ 7 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=8 + [[ 8 -eq 0 ]] + [[ 8 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=9 + [[ 9 -eq 0 ]] + [[ 9 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result='{"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"}' + [[ {"additional_password": "$A$005$\\u0001EUjJ2#\\u000fySR?)w\\r+d=t\\u000439NBeGFTfqMqRaBY2.AyAvoGohhnbklzmw6y3wTolwB"} =~ additional_password ]] + return 1 + sleep 1 + retry=10 + [[ 0 -eq 0 ]] + [[ 10 -gt 0 ]] + echo 'waiting for old password to be discarded... (retry 10/600)' waiting for old password to be discarded... (retry 10/600) + [[ 10 -ge 600 ]] + is_old_password_discarded monitor '-h some-name-pxc -uroot -p'\''test-password'\''' + local username=monitor + local 'uri=-h some-name-pxc -uroot -p'\''test-password'\''' + local result ++ run_mysql 'SELECT User_attributes FROM mysql.user WHERE user='\''monitor'\''' '-h some-name-pxc -uroot -p'\''test-password'\''' + result=NULL + [[ NULL =~ additional_password ]] + [[ -z NULL ]] + [[ NULL =~ ^[[:space:]]*$ ]] + [[ NULL =~ NULL ]] + return 0 + echo 'Old password successfully discarded after 10 retries' Old password successfully discarded after 10 retries + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dWKzp9Yuw4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.vH9ES9SaQ8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dWKzp9Yuw4 ++ cat /tmp/tmp.vH9ES9SaQ8 ++ rm /tmp/tmp.dWKzp9Yuw4 /tmp/tmp.vH9ES9SaQ8 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jqmCVfaKEt +++ mktemp ++ local LAST_ERR=/tmp/tmp.yNVRTkksfF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jqmCVfaKEt ++ cat /tmp/tmp.yNVRTkksfF ++ rm /tmp/tmp.jqmCVfaKEt /tmp/tmp.yNVRTkksfF ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.2gtdA8Dstd ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.WkxOeH3Zxg +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.2gtdA8Dstd +++++ cat /tmp/tmp.WkxOeH3Zxg +++++ rm /tmp/tmp.2gtdA8Dstd /tmp/tmp.WkxOeH3Zxg +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.VaeVnKSaB9 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.0M0ONzYpCT +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.VaeVnKSaB9 +++++ cat /tmp/tmp.0M0ONzYpCT +++++ rm /tmp/tmp.VaeVnKSaB9 /tmp/tmp.0M0ONzYpCT +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c9NI5pgxNN +++ mktemp ++ local LAST_ERR=/tmp/tmp.jq3Fmfcyk6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.c9NI5pgxNN ++ cat /tmp/tmp.jq3Fmfcyk6 ++ rm /tmp/tmp.c9NI5pgxNN /tmp/tmp.jq3Fmfcyk6 ++ return 0 + [[ 3 == \2 ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ov56FUxq0s +++ mktemp ++ local LAST_ERR=/tmp/tmp.wIJWD9vv8E ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ov56FUxq0s ++ cat /tmp/tmp.wIJWD9vv8E ++ rm /tmp/tmp.ov56FUxq0s /tmp/tmp.wIJWD9vv8E ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6qpJrBWstD +++ mktemp ++ local LAST_ERR=/tmp/tmp.9uJDG2lXQs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6qpJrBWstD ++ cat /tmp/tmp.9uJDG2lXQs ++ rm /tmp/tmp.6qpJrBWstD /tmp/tmp.9uJDG2lXQs ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.NZA5MpOEnd ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.latD1iiZEu +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.NZA5MpOEnd +++++ cat /tmp/tmp.latD1iiZEu +++++ rm /tmp/tmp.NZA5MpOEnd /tmp/tmp.latD1iiZEu +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.XcRujyod4l ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.hWXuCh1Jnw +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.XcRujyod4l +++++ cat /tmp/tmp.hWXuCh1Jnw +++++ rm /tmp/tmp.XcRujyod4l /tmp/tmp.hWXuCh1Jnw +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eZqgPnLGCO +++ mktemp ++ local LAST_ERR=/tmp/tmp.OZIwaMSnGu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eZqgPnLGCO ++ cat /tmp/tmp.OZIwaMSnGu ++ rm /tmp/tmp.eZqgPnLGCO /tmp/tmp.OZIwaMSnGu ++ return 0 + [[ 3 == \2 ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v04PU0GuN2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0XgbKqE0nj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v04PU0GuN2 ++ cat /tmp/tmp.0XgbKqE0nj ++ rm /tmp/tmp.v04PU0GuN2 /tmp/tmp.0XgbKqE0nj ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IEvxYG6TJR +++ mktemp ++ local LAST_ERR=/tmp/tmp.SPDIFy7a0p ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IEvxYG6TJR ++ cat /tmp/tmp.SPDIFy7a0p ++ rm /tmp/tmp.IEvxYG6TJR /tmp/tmp.SPDIFy7a0p ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.r6ExBaDGgw ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.KruuosYysK +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.r6ExBaDGgw +++++ cat /tmp/tmp.KruuosYysK +++++ rm /tmp/tmp.r6ExBaDGgw /tmp/tmp.KruuosYysK +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.0CYl3tLQEI ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.TtrQSyn2nG +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.0CYl3tLQEI +++++ cat /tmp/tmp.TtrQSyn2nG +++++ rm /tmp/tmp.0CYl3tLQEI /tmp/tmp.TtrQSyn2nG +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DZFxzSQGrI +++ mktemp ++ local LAST_ERR=/tmp/tmp.j4rKJCZaWq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DZFxzSQGrI ++ cat /tmp/tmp.j4rKJCZaWq ++ rm /tmp/tmp.DZFxzSQGrI /tmp/tmp.j4rKJCZaWq ++ return 0 + [[ 3 == \2 ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.T5iRFjSJPC +++ mktemp ++ local LAST_ERR=/tmp/tmp.T21WWbSCBE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.T5iRFjSJPC ++ cat /tmp/tmp.T21WWbSCBE ++ rm /tmp/tmp.T5iRFjSJPC /tmp/tmp.T21WWbSCBE ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zFcT5CX1FN +++ mktemp ++ local LAST_ERR=/tmp/tmp.MmOgCnGvk3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zFcT5CX1FN ++ cat /tmp/tmp.MmOgCnGvk3 ++ rm /tmp/tmp.zFcT5CX1FN /tmp/tmp.MmOgCnGvk3 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.fzvQJJ2TE3 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.vDqSXV7L9y +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.fzvQJJ2TE3 +++++ cat /tmp/tmp.vDqSXV7L9y +++++ rm /tmp/tmp.fzvQJJ2TE3 /tmp/tmp.vDqSXV7L9y +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.4CHJ5QbSff ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.zMffpXVy6f +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.4CHJ5QbSff +++++ cat /tmp/tmp.zMffpXVy6f +++++ rm /tmp/tmp.4CHJ5QbSff /tmp/tmp.zMffpXVy6f +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lAuDSMHaXy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Op6acvS7ah ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lAuDSMHaXy ++ cat /tmp/tmp.Op6acvS7ah ++ rm /tmp/tmp.lAuDSMHaXy /tmp/tmp.Op6acvS7ah ++ return 0 + [[ 3 == \2 ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w54M8AJCdA +++ mktemp ++ local LAST_ERR=/tmp/tmp.jm2czAJsFe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.w54M8AJCdA ++ cat /tmp/tmp.jm2czAJsFe ++ rm /tmp/tmp.w54M8AJCdA /tmp/tmp.jm2czAJsFe ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vuVFrKsuvV +++ mktemp ++ local LAST_ERR=/tmp/tmp.lmALNRLuIN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vuVFrKsuvV ++ cat /tmp/tmp.lmALNRLuIN ++ rm /tmp/tmp.vuVFrKsuvV /tmp/tmp.lmALNRLuIN ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.FNXxIAHVFu ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.FOLjpMCbQX +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.FNXxIAHVFu +++++ cat /tmp/tmp.FOLjpMCbQX +++++ rm /tmp/tmp.FNXxIAHVFu /tmp/tmp.FOLjpMCbQX +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.xAzRF5JEBv ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.vtzpCkJ60l +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.xAzRF5JEBv +++++ cat /tmp/tmp.vtzpCkJ60l +++++ rm /tmp/tmp.xAzRF5JEBv /tmp/tmp.vtzpCkJ60l +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oKP2ANKtK3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3BDhRQUh9a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oKP2ANKtK3 ++ cat /tmp/tmp.3BDhRQUh9a ++ rm /tmp/tmp.oKP2ANKtK3 /tmp/tmp.3BDhRQUh9a ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -umonitor -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -umonitor -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vMdMonXLsq +++ mktemp ++ local LAST_ERR=/tmp/tmp.cXVz7AcD35 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vMdMonXLsq ++ cat /tmp/tmp.cXVz7AcD35 ++ rm /tmp/tmp.vMdMonXLsq /tmp/tmp.cXVz7AcD35 ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + compare_mysql_cmd_local proxysql-cfg 'SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 '' proxysql + local command_id=proxysql-cfg + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local postfix= + local container_name=proxysql + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/proxysql-cfg.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/proxysql-cfg-84.sql ']' + run_mysql_local 'SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' some-name-proxysql-1 proxysql + local 'command=SELECT variable_value from global_variables WHERE variable_name='\''mysql-monitor_password'\'';' + local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -p'\''test-password'\''' + local pod=some-name-proxysql-1 + local container_name=proxysql + set +o xtrace + '[' '!' -s /tmp/tmp.uv48lQ0yvW/proxysql-cfg.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/proxysql-cfg.sql /tmp/tmp.uv48lQ0yvW/proxysql-cfg.sql + desc 'test operator' + set +o xtrace ----------------------------------------------------------------------------------- test operator ----------------------------------------------------------------------------------- + patch_secret my-cluster-secrets operator dGVzdC1wYXNzd29yZA== + local secret=my-cluster-secrets + local key=operator + local value=dGVzdC1wYXNzd29yZA== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.ch7KIcSV3y ++ mktemp + local LAST_ERR=/tmp/tmp.jGyx6h0hjF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"operator": "dGVzdC1wYXNzd29yZA=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ch7KIcSV3y secret/my-cluster-secrets patched + cat /tmp/tmp.jGyx6h0hjF + rm /tmp/tmp.ch7KIcSV3y /tmp/tmp.jGyx6h0hjF + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LXrqzJvfBS +++ mktemp ++ local LAST_ERR=/tmp/tmp.tOSkl9F0fE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LXrqzJvfBS ++ cat /tmp/tmp.tOSkl9F0fE ++ rm /tmp/tmp.LXrqzJvfBS /tmp/tmp.tOSkl9F0fE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XCb3ejtOtm +++ mktemp ++ local LAST_ERR=/tmp/tmp.4EsuNEVun3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XCb3ejtOtm ++ cat /tmp/tmp.4EsuNEVun3 ++ rm /tmp/tmp.XCb3ejtOtm /tmp/tmp.4EsuNEVun3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fI1yo5QZTL +++ mktemp ++ local LAST_ERR=/tmp/tmp.bHgBHfekwg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fI1yo5QZTL ++ cat /tmp/tmp.bHgBHfekwg ++ rm /tmp/tmp.fI1yo5QZTL /tmp/tmp.bHgBHfekwg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BCIiGGPaaE +++ mktemp ++ local LAST_ERR=/tmp/tmp.51DdMw2EH5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BCIiGGPaaE ++ cat /tmp/tmp.51DdMw2EH5 ++ rm /tmp/tmp.BCIiGGPaaE /tmp/tmp.51DdMw2EH5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xFyOw0C8ON +++ mktemp ++ local LAST_ERR=/tmp/tmp.dEpJVkH8gX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xFyOw0C8ON ++ cat /tmp/tmp.dEpJVkH8gX ++ rm /tmp/tmp.xFyOw0C8ON /tmp/tmp.dEpJVkH8gX ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uQEdcQJ0Sg +++ mktemp ++ local LAST_ERR=/tmp/tmp.Kl4uDg0vpX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uQEdcQJ0Sg ++ cat /tmp/tmp.Kl4uDg0vpX ++ rm /tmp/tmp.uQEdcQJ0Sg /tmp/tmp.Kl4uDg0vpX ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.TltoTIQxyJ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.yCmQOPLE6P +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.TltoTIQxyJ +++++ cat /tmp/tmp.yCmQOPLE6P +++++ rm /tmp/tmp.TltoTIQxyJ /tmp/tmp.yCmQOPLE6P +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.9dkIPbSW1W ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.q0kErikwxm +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.9dkIPbSW1W +++++ cat /tmp/tmp.q0kErikwxm +++++ rm /tmp/tmp.9dkIPbSW1W /tmp/tmp.q0kErikwxm +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SE66frR7WQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z0FaojIDac ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SE66frR7WQ ++ cat /tmp/tmp.Z0FaojIDac ++ rm /tmp/tmp.SE66frR7WQ /tmp/tmp.Z0FaojIDac ++ return 0 + [[ 2 == \2 ]] + echo + sleep 15 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rHoFCd83ob +++ mktemp ++ local LAST_ERR=/tmp/tmp.zd8CqLdiJH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rHoFCd83ob ++ cat /tmp/tmp.zd8CqLdiJH ++ rm /tmp/tmp.rHoFCd83ob /tmp/tmp.zd8CqLdiJH ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + desc 'change secret name' + set +o xtrace ----------------------------------------------------------------------------------- change secret name ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.NuzGbSOFRh ++ mktemp + local LAST_ERR=/tmp/tmp.IRDCpYOgp1 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type merge --patch '{"spec": {"secretsName":"my-cluster-secrets-2"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NuzGbSOFRh perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.IRDCpYOgp1 + rm /tmp/tmp.NuzGbSOFRh /tmp/tmp.IRDCpYOgp1 + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P06IMSgzwB +++ mktemp ++ local LAST_ERR=/tmp/tmp.iSpOYu6ErN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.P06IMSgzwB ++ cat /tmp/tmp.iSpOYu6ErN ++ rm /tmp/tmp.P06IMSgzwB /tmp/tmp.iSpOYu6ErN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c06fYu9PAs +++ mktemp ++ local LAST_ERR=/tmp/tmp.beAty1vjjC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.c06fYu9PAs ++ cat /tmp/tmp.beAty1vjjC ++ rm /tmp/tmp.c06fYu9PAs /tmp/tmp.beAty1vjjC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BPTPwXqQAE +++ mktemp ++ local LAST_ERR=/tmp/tmp.PxYC0cxh0K ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BPTPwXqQAE ++ cat /tmp/tmp.PxYC0cxh0K ++ rm /tmp/tmp.BPTPwXqQAE /tmp/tmp.PxYC0cxh0K ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PAI5F4OuSC +++ mktemp ++ local LAST_ERR=/tmp/tmp.eUy3HkCNev ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PAI5F4OuSC ++ cat /tmp/tmp.eUy3HkCNev ++ rm /tmp/tmp.PAI5F4OuSC /tmp/tmp.eUy3HkCNev ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ylLm2VNRqU +++ mktemp ++ local LAST_ERR=/tmp/tmp.Yri3TDTl2U ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ylLm2VNRqU ++ cat /tmp/tmp.Yri3TDTl2U ++ rm /tmp/tmp.ylLm2VNRqU /tmp/tmp.Yri3TDTl2U ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.foxwmK9Yoi +++ mktemp ++ local LAST_ERR=/tmp/tmp.MK0xjp0gDP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.foxwmK9Yoi ++ cat /tmp/tmp.MK0xjp0gDP ++ rm /tmp/tmp.foxwmK9Yoi /tmp/tmp.MK0xjp0gDP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sZdrfAGhuw +++ mktemp ++ local LAST_ERR=/tmp/tmp.8IL9Y2lEYV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sZdrfAGhuw ++ cat /tmp/tmp.8IL9Y2lEYV ++ rm /tmp/tmp.sZdrfAGhuw /tmp/tmp.8IL9Y2lEYV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U5sYWx7OZI +++ mktemp ++ local LAST_ERR=/tmp/tmp.rTmkZs3Fex ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.U5sYWx7OZI ++ cat /tmp/tmp.rTmkZs3Fex ++ rm /tmp/tmp.U5sYWx7OZI /tmp/tmp.rTmkZs3Fex ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ruPS8q8BL1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Gwn5Ik4kMK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ruPS8q8BL1 ++ cat /tmp/tmp.Gwn5Ik4kMK ++ rm /tmp/tmp.ruPS8q8BL1 /tmp/tmp.Gwn5Ik4kMK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v6C68GwHpy +++ mktemp ++ local LAST_ERR=/tmp/tmp.bvfC4d9gth ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v6C68GwHpy ++ cat /tmp/tmp.bvfC4d9gth ++ rm /tmp/tmp.v6C68GwHpy /tmp/tmp.bvfC4d9gth ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xXsfWw5QSY +++ mktemp ++ local LAST_ERR=/tmp/tmp.lpESLL5JLY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xXsfWw5QSY ++ cat /tmp/tmp.lpESLL5JLY ++ rm /tmp/tmp.xXsfWw5QSY /tmp/tmp.lpESLL5JLY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HWzAPcvIhX +++ mktemp ++ local LAST_ERR=/tmp/tmp.TbVhgJ71O4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HWzAPcvIhX ++ cat /tmp/tmp.TbVhgJ71O4 ++ rm /tmp/tmp.HWzAPcvIhX /tmp/tmp.TbVhgJ71O4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zYaICRJnV7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.4NML1RNURM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zYaICRJnV7 ++ cat /tmp/tmp.4NML1RNURM ++ rm /tmp/tmp.zYaICRJnV7 /tmp/tmp.4NML1RNURM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7pIXPJRoxU +++ mktemp ++ local LAST_ERR=/tmp/tmp.k3nhKq0FXq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7pIXPJRoxU ++ cat /tmp/tmp.k3nhKq0FXq ++ rm /tmp/tmp.7pIXPJRoxU /tmp/tmp.k3nhKq0FXq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nMER9vLAvS +++ mktemp ++ local LAST_ERR=/tmp/tmp.kdCq1ndueD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nMER9vLAvS ++ cat /tmp/tmp.kdCq1ndueD ++ rm /tmp/tmp.nMER9vLAvS /tmp/tmp.kdCq1ndueD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9p71pisdgF +++ mktemp ++ local LAST_ERR=/tmp/tmp.ptfL0BcXcm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9p71pisdgF ++ cat /tmp/tmp.ptfL0BcXcm ++ rm /tmp/tmp.9p71pisdgF /tmp/tmp.ptfL0BcXcm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.W12mQPvNic +++ mktemp ++ local LAST_ERR=/tmp/tmp.A9V95O0YFx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.W12mQPvNic ++ cat /tmp/tmp.A9V95O0YFx ++ rm /tmp/tmp.W12mQPvNic /tmp/tmp.A9V95O0YFx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z6KHQRp6lF +++ mktemp ++ local LAST_ERR=/tmp/tmp.oFxnvLYSb3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z6KHQRp6lF ++ cat /tmp/tmp.oFxnvLYSb3 ++ rm /tmp/tmp.z6KHQRp6lF /tmp/tmp.oFxnvLYSb3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5cBIWSDxEL +++ mktemp ++ local LAST_ERR=/tmp/tmp.kWsDleWWFH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5cBIWSDxEL ++ cat /tmp/tmp.kWsDleWWFH ++ rm /tmp/tmp.5cBIWSDxEL /tmp/tmp.kWsDleWWFH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oSDL5FWzOb +++ mktemp ++ local LAST_ERR=/tmp/tmp.n32uBCWcfy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oSDL5FWzOb ++ cat /tmp/tmp.n32uBCWcfy ++ rm /tmp/tmp.oSDL5FWzOb /tmp/tmp.n32uBCWcfy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q8fTZPLSFP +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ka60CMqP2P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.q8fTZPLSFP ++ cat /tmp/tmp.Ka60CMqP2P ++ rm /tmp/tmp.q8fTZPLSFP /tmp/tmp.Ka60CMqP2P ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PhFXolpWgk +++ mktemp ++ local LAST_ERR=/tmp/tmp.ib0ALC5Wzp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PhFXolpWgk ++ cat /tmp/tmp.ib0ALC5Wzp ++ rm /tmp/tmp.PhFXolpWgk /tmp/tmp.ib0ALC5Wzp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.meeNXC3Q7G +++ mktemp ++ local LAST_ERR=/tmp/tmp.FOx8FlD6ZP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.meeNXC3Q7G ++ cat /tmp/tmp.FOx8FlD6ZP ++ rm /tmp/tmp.meeNXC3Q7G /tmp/tmp.FOx8FlD6ZP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PzEMeH7YmH +++ mktemp ++ local LAST_ERR=/tmp/tmp.M9Mv1kq55P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PzEMeH7YmH ++ cat /tmp/tmp.M9Mv1kq55P ++ rm /tmp/tmp.PzEMeH7YmH /tmp/tmp.M9Mv1kq55P ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7f4XIP8YIn +++ mktemp ++ local LAST_ERR=/tmp/tmp.FNIIUjqkBr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7f4XIP8YIn ++ cat /tmp/tmp.FNIIUjqkBr ++ rm /tmp/tmp.7f4XIP8YIn /tmp/tmp.FNIIUjqkBr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zovSqeg5cS +++ mktemp ++ local LAST_ERR=/tmp/tmp.B2x0DsTbcK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zovSqeg5cS ++ cat /tmp/tmp.B2x0DsTbcK ++ rm /tmp/tmp.zovSqeg5cS /tmp/tmp.B2x0DsTbcK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MKdTlOA8qt +++ mktemp ++ local LAST_ERR=/tmp/tmp.ycewnpmlCE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MKdTlOA8qt ++ cat /tmp/tmp.ycewnpmlCE ++ rm /tmp/tmp.MKdTlOA8qt /tmp/tmp.ycewnpmlCE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 26 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VTyVTmqORY +++ mktemp ++ local LAST_ERR=/tmp/tmp.MRkNL0UK8n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VTyVTmqORY ++ cat /tmp/tmp.MRkNL0UK8n ++ rm /tmp/tmp.VTyVTmqORY /tmp/tmp.MRkNL0UK8n ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 27 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NeTRiPwfzn +++ mktemp ++ local LAST_ERR=/tmp/tmp.Th8u6aUvdY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NeTRiPwfzn ++ cat /tmp/tmp.Th8u6aUvdY ++ rm /tmp/tmp.NeTRiPwfzn /tmp/tmp.Th8u6aUvdY ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NkTCPAWkQ7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.e27brCjpc4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NkTCPAWkQ7 ++ cat /tmp/tmp.e27brCjpc4 ++ rm /tmp/tmp.NkTCPAWkQ7 /tmp/tmp.e27brCjpc4 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.vHfoQskLrk ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.fYpRaMIGIZ +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.vHfoQskLrk +++++ cat /tmp/tmp.fYpRaMIGIZ +++++ rm /tmp/tmp.vHfoQskLrk /tmp/tmp.fYpRaMIGIZ +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Y83kFhrLUL ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.09l81bJHEf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Y83kFhrLUL +++++ cat /tmp/tmp.09l81bJHEf +++++ rm /tmp/tmp.Y83kFhrLUL /tmp/tmp.09l81bJHEf +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.niJa9D5Ctq +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZsxbT4PrY5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.niJa9D5Ctq ++ cat /tmp/tmp.ZsxbT4PrY5 ++ rm /tmp/tmp.niJa9D5Ctq /tmp/tmp.ZsxbT4PrY5 ++ return 0 + [[ 2 == \2 ]] + echo + desc 'test new operator' + set +o xtrace ----------------------------------------------------------------------------------- test new operator ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + patch_secret my-cluster-secrets-2 operator dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets-2 + local key=operator + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.ZvoZ35TLR7 ++ mktemp + local LAST_ERR=/tmp/tmp.fD1GCAkdwk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets-2 '-p={"data":{"operator": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZvoZ35TLR7 secret/my-cluster-secrets-2 patched + cat /tmp/tmp.fD1GCAkdwk + rm /tmp/tmp.ZvoZ35TLR7 /tmp/tmp.fD1GCAkdwk + return 0 + sleep 15 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Yj6wIGI33J +++ mktemp ++ local LAST_ERR=/tmp/tmp.CCAz4INapT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Yj6wIGI33J ++ cat /tmp/tmp.CCAz4INapT ++ rm /tmp/tmp.Yj6wIGI33J /tmp/tmp.CCAz4INapT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HaaQbg1N5C +++ mktemp ++ local LAST_ERR=/tmp/tmp.e3vRG0Qn70 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HaaQbg1N5C ++ cat /tmp/tmp.e3vRG0Qn70 ++ rm /tmp/tmp.HaaQbg1N5C /tmp/tmp.e3vRG0Qn70 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.amQmS6WyUI +++ mktemp ++ local LAST_ERR=/tmp/tmp.f4m13GNu0M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.amQmS6WyUI ++ cat /tmp/tmp.f4m13GNu0M ++ rm /tmp/tmp.amQmS6WyUI /tmp/tmp.f4m13GNu0M ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y9b4vdlDeZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ft380eQQZP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.y9b4vdlDeZ ++ cat /tmp/tmp.Ft380eQQZP ++ rm /tmp/tmp.y9b4vdlDeZ /tmp/tmp.Ft380eQQZP ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5hQGyXkkNG +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dc6WWmMvSv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5hQGyXkkNG ++ cat /tmp/tmp.Dc6WWmMvSv ++ rm /tmp/tmp.5hQGyXkkNG /tmp/tmp.Dc6WWmMvSv ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.IwW3pkgdCP ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.cvK5Um3NWP +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.IwW3pkgdCP +++++ cat /tmp/tmp.cvK5Um3NWP +++++ rm /tmp/tmp.IwW3pkgdCP /tmp/tmp.cvK5Um3NWP +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.cvqouhgViB ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.vq4lqbAx7y +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.cvqouhgViB +++++ cat /tmp/tmp.vq4lqbAx7y +++++ rm /tmp/tmp.cvqouhgViB /tmp/tmp.vq4lqbAx7y +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2xPkzAFf0H +++ mktemp ++ local LAST_ERR=/tmp/tmp.xfTbm6qwcN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2xPkzAFf0H ++ cat /tmp/tmp.xfTbm6qwcN ++ rm /tmp/tmp.2xPkzAFf0H /tmp/tmp.xfTbm6qwcN ++ return 0 + [[ 2 == \2 ]] + echo + sleep 20 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bjLl91eetP +++ mktemp ++ local LAST_ERR=/tmp/tmp.b5mgE6UCts ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bjLl91eetP ++ cat /tmp/tmp.b5mgE6UCts ++ rm /tmp/tmp.bjLl91eetP /tmp/tmp.b5mgE6UCts ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql ++ getSecretData my-cluster-secrets-2 root ++ local secretName=my-cluster-secrets-2 ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.1b66lD6sEM +++ mktemp ++ local LAST_ERR=/tmp/tmp.bPR4AGww9W ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets-2 '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1b66lD6sEM ++ cat /tmp/tmp.bPR4AGww9W ++ rm /tmp/tmp.1b66lD6sEM /tmp/tmp.bPR4AGww9W ++ return 0 + newpass='K!m2SQOes@>G[?iBa' + desc 'test new users sync' + set +o xtrace ----------------------------------------------------------------------------------- test new users sync ----------------------------------------------------------------------------------- + run_mysql 'CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''K!m2SQOes@>G[?iBa'\'';' '-h some-name-pxc -uroot -p'\''K!m2SQOes@>G[?iBa'\''' + local 'command=CREATE USER '\''testsync'\''@'\''%'\'' IDENTIFIED BY '\''K!m2SQOes@>G[?iBa'\'';' + local 'uri=-h some-name-pxc -uroot -p'\''K!m2SQOes@>G[?iBa'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FEx6HqKVM5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.p4W9sneb7M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FEx6HqKVM5 ++ cat /tmp/tmp.p4W9sneb7M ++ rm /tmp/tmp.FEx6HqKVM5 /tmp/tmp.p4W9sneb7M ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 80 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''K!m2SQOes@>G[?iBa'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''K!m2SQOes@>G[?iBa'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -utestsync -p'\''K!m2SQOes@>G[?iBa'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -utestsync -p'\''K!m2SQOes@>G[?iBa'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.j9QOqfeLk0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.BrGHDtzQNX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.j9QOqfeLk0 ++ cat /tmp/tmp.BrGHDtzQNX ++ rm /tmp/tmp.j9QOqfeLk0 /tmp/tmp.BrGHDtzQNX ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql ++ getSecretData internal-some-name operator ++ local secretName=internal-some-name ++ local dataKey=operator ++ kubectl_bin get secrets/internal-some-name '--template={{.data.operator}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.dVXJpvcGFy +++ mktemp ++ local LAST_ERR=/tmp/tmp.iIzh9BoQil ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/internal-some-name '--template={{.data.operator}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dVXJpvcGFy ++ cat /tmp/tmp.iIzh9BoQil ++ rm /tmp/tmp.dVXJpvcGFy /tmp/tmp.iIzh9BoQil ++ return 0 + pass=test-password2 + desc 'check secret without operator' + set +o xtrace ----------------------------------------------------------------------------------- check secret without operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9em4Pb77EO ++ mktemp + local LAST_ERR=/tmp/tmp.Fj3hgEG0ye + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9em4Pb77EO secret/my-cluster-secrets-2 configured + cat /tmp/tmp.Fj3hgEG0ye Warning: resource secrets/my-cluster-secrets-2 is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.9em4Pb77EO /tmp/tmp.Fj3hgEG0ye + return 0 + sleep 60 + compare_mysql_cmd select-4 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local command_id=select-4 + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql + run_mysql 'SHOW TABLES;' '-h some-name-proxysql -uoperator -p'\''test-password2'\''' + local 'command=SHOW TABLES;' + local 'uri=-h some-name-proxysql -uoperator -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lPo1W6EaYJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.a9NXwZ1o5N ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lPo1W6EaYJ ++ cat /tmp/tmp.a9NXwZ1o5N ++ rm /tmp/tmp.lPo1W6EaYJ /tmp/tmp.a9NXwZ1o5N ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-4.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-4.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-4-84.sql /tmp/tmp.uv48lQ0yvW/select-4.sql + desc 'test enable haproxy and disable proxysql ' + set +o xtrace ----------------------------------------------------------------------------------- test enable haproxy and disable proxysql ----------------------------------------------------------------------------------- + newpass=test-password2 ++ echo -n test-password2 ++ base64 + newpassencrypted=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch pxc some-name --type=merge -p '{"spec":{"haproxy":{"enabled":true},"proxysql":{"enabled":false},"secretsName":"my-cluster-secrets"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.CHxpdZq2zw ++ mktemp + local LAST_ERR=/tmp/tmp.Si9IM74g35 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge -p '{"spec":{"haproxy":{"enabled":true},"proxysql":{"enabled":false},"secretsName":"my-cluster-secrets"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CHxpdZq2zw perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.Si9IM74g35 + rm /tmp/tmp.CHxpdZq2zw /tmp/tmp.Si9IM74g35 + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dhpmMV5Yw5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.l8OHnSQHcG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dhpmMV5Yw5 ++ cat /tmp/tmp.l8OHnSQHcG ++ rm /tmp/tmp.dhpmMV5Yw5 /tmp/tmp.l8OHnSQHcG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uySKao2R2Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.cxr4qqdfWL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uySKao2R2Q ++ cat /tmp/tmp.cxr4qqdfWL ++ rm /tmp/tmp.uySKao2R2Q /tmp/tmp.cxr4qqdfWL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JTly84LEZN +++ mktemp ++ local LAST_ERR=/tmp/tmp.2ekdYLcCkv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JTly84LEZN ++ cat /tmp/tmp.2ekdYLcCkv ++ rm /tmp/tmp.JTly84LEZN /tmp/tmp.2ekdYLcCkv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DaDd2khrmF +++ mktemp ++ local LAST_ERR=/tmp/tmp.SXVyF9HQcT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DaDd2khrmF ++ cat /tmp/tmp.SXVyF9HQcT ++ rm /tmp/tmp.DaDd2khrmF /tmp/tmp.SXVyF9HQcT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QZYWnAcemw +++ mktemp ++ local LAST_ERR=/tmp/tmp.BvbbktWJNm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QZYWnAcemw ++ cat /tmp/tmp.BvbbktWJNm ++ rm /tmp/tmp.QZYWnAcemw /tmp/tmp.BvbbktWJNm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IDSXors3Ya +++ mktemp ++ local LAST_ERR=/tmp/tmp.zfxLL8RqLU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IDSXors3Ya ++ cat /tmp/tmp.zfxLL8RqLU ++ rm /tmp/tmp.IDSXors3Ya /tmp/tmp.zfxLL8RqLU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dThAgGxsNm +++ mktemp ++ local LAST_ERR=/tmp/tmp.SuDKxpkdUE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dThAgGxsNm ++ cat /tmp/tmp.SuDKxpkdUE ++ rm /tmp/tmp.dThAgGxsNm /tmp/tmp.SuDKxpkdUE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tLpmjfEdC3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.i02CMlM8Kc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tLpmjfEdC3 ++ cat /tmp/tmp.i02CMlM8Kc ++ rm /tmp/tmp.tLpmjfEdC3 /tmp/tmp.i02CMlM8Kc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gf4B6N9f61 +++ mktemp ++ local LAST_ERR=/tmp/tmp.apON3NfgQ3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gf4B6N9f61 ++ cat /tmp/tmp.apON3NfgQ3 ++ rm /tmp/tmp.gf4B6N9f61 /tmp/tmp.apON3NfgQ3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.haT8dOkcNG +++ mktemp ++ local LAST_ERR=/tmp/tmp.Gf4eUify1l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.haT8dOkcNG ++ cat /tmp/tmp.Gf4eUify1l ++ rm /tmp/tmp.haT8dOkcNG /tmp/tmp.Gf4eUify1l ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ulwhcZZYSs +++ mktemp ++ local LAST_ERR=/tmp/tmp.KPMOvgZWol ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ulwhcZZYSs ++ cat /tmp/tmp.KPMOvgZWol ++ rm /tmp/tmp.ulwhcZZYSs /tmp/tmp.KPMOvgZWol ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AzBN2aRkV1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.BOmmSH7j0A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AzBN2aRkV1 ++ cat /tmp/tmp.BOmmSH7j0A ++ rm /tmp/tmp.AzBN2aRkV1 /tmp/tmp.BOmmSH7j0A ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FQMabCiQOx +++ mktemp ++ local LAST_ERR=/tmp/tmp.jWExAY3rGk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FQMabCiQOx ++ cat /tmp/tmp.jWExAY3rGk ++ rm /tmp/tmp.FQMabCiQOx /tmp/tmp.jWExAY3rGk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H9XwlwyzlS +++ mktemp ++ local LAST_ERR=/tmp/tmp.3MQd3vUneX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.H9XwlwyzlS ++ cat /tmp/tmp.3MQd3vUneX ++ rm /tmp/tmp.H9XwlwyzlS /tmp/tmp.3MQd3vUneX ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wBYHRAxm7l +++ mktemp ++ local LAST_ERR=/tmp/tmp.PYukiqHLRq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wBYHRAxm7l ++ cat /tmp/tmp.PYukiqHLRq ++ rm /tmp/tmp.wBYHRAxm7l /tmp/tmp.PYukiqHLRq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DZCCK952Qq +++ mktemp ++ local LAST_ERR=/tmp/tmp.eZMyrsvC1E ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DZCCK952Qq ++ cat /tmp/tmp.eZMyrsvC1E ++ rm /tmp/tmp.DZCCK952Qq /tmp/tmp.eZMyrsvC1E ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CFbsBqYm1y +++ mktemp ++ local LAST_ERR=/tmp/tmp.puOVeivStA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CFbsBqYm1y ++ cat /tmp/tmp.puOVeivStA ++ rm /tmp/tmp.CFbsBqYm1y /tmp/tmp.puOVeivStA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.76R7mGSKvX +++ mktemp ++ local LAST_ERR=/tmp/tmp.NHgMewA6oa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.76R7mGSKvX ++ cat /tmp/tmp.NHgMewA6oa ++ rm /tmp/tmp.76R7mGSKvX /tmp/tmp.NHgMewA6oa ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nte1O2orZ8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.GMxwvNMtbx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nte1O2orZ8 ++ cat /tmp/tmp.GMxwvNMtbx ++ rm /tmp/tmp.nte1O2orZ8 /tmp/tmp.GMxwvNMtbx ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.LxlcqTAhjt ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Ok6spvPgaj +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.LxlcqTAhjt +++++ cat /tmp/tmp.Ok6spvPgaj +++++ rm /tmp/tmp.LxlcqTAhjt /tmp/tmp.Ok6spvPgaj +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JPqDrdXQ7N +++ mktemp ++ local LAST_ERR=/tmp/tmp.NE2oRZoqoe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JPqDrdXQ7N ++ cat /tmp/tmp.NE2oRZoqoe ++ rm /tmp/tmp.JPqDrdXQ7N /tmp/tmp.NE2oRZoqoe ++ return 0 + [[ 3 == \3 ]] + echo + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + check_generation 1 haproxy some-name + local generation=1 + local container=haproxy + local cluster=some-name + local current_generation ++ kubectl_bin get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c0s2mBGuiy +++ mktemp ++ local LAST_ERR=/tmp/tmp.drdLUZ3wik ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.c0s2mBGuiy ++ cat /tmp/tmp.drdLUZ3wik ++ rm /tmp/tmp.c0s2mBGuiy /tmp/tmp.drdLUZ3wik ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + patch_secret my-cluster-secrets monitor dGVzdC1wYXNzd29yZDI= + local secret=my-cluster-secrets + local key=monitor + local value=dGVzdC1wYXNzd29yZDI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.D8fn3QRJyx ++ mktemp + local LAST_ERR=/tmp/tmp.14hSawlbsp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"monitor": "dGVzdC1wYXNzd29yZDI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.D8fn3QRJyx secret/my-cluster-secrets patched + cat /tmp/tmp.14hSawlbsp + rm /tmp/tmp.D8fn3QRJyx /tmp/tmp.14hSawlbsp + return 0 + sleep 15 + wait_cluster_consistency some-name 3 3 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=3 + '[' -z 3 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pGWcsnICkn +++ mktemp ++ local LAST_ERR=/tmp/tmp.BlU98XCaWk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pGWcsnICkn ++ cat /tmp/tmp.BlU98XCaWk ++ rm /tmp/tmp.pGWcsnICkn /tmp/tmp.BlU98XCaWk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QnCSIen7ah +++ mktemp ++ local LAST_ERR=/tmp/tmp.DX28VbkLUN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QnCSIen7ah ++ cat /tmp/tmp.DX28VbkLUN ++ rm /tmp/tmp.QnCSIen7ah /tmp/tmp.DX28VbkLUN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ugxckfiPK3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.aqGd0GnTAd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ugxckfiPK3 ++ cat /tmp/tmp.aqGd0GnTAd ++ rm /tmp/tmp.ugxckfiPK3 /tmp/tmp.aqGd0GnTAd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.G0no5uQbra +++ mktemp ++ local LAST_ERR=/tmp/tmp.EWB9U9DqkT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.G0no5uQbra ++ cat /tmp/tmp.EWB9U9DqkT ++ rm /tmp/tmp.G0no5uQbra /tmp/tmp.EWB9U9DqkT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nREIv6sN0n +++ mktemp ++ local LAST_ERR=/tmp/tmp.OrwgeMkJfF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nREIv6sN0n ++ cat /tmp/tmp.OrwgeMkJfF ++ rm /tmp/tmp.nREIv6sN0n /tmp/tmp.OrwgeMkJfF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.J1pyKKSUVc +++ mktemp ++ local LAST_ERR=/tmp/tmp.fp1YpDWNG1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.J1pyKKSUVc ++ cat /tmp/tmp.fp1YpDWNG1 ++ rm /tmp/tmp.J1pyKKSUVc /tmp/tmp.fp1YpDWNG1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ntKjDQB42C +++ mktemp ++ local LAST_ERR=/tmp/tmp.rjsIMScLNA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ntKjDQB42C ++ cat /tmp/tmp.rjsIMScLNA ++ rm /tmp/tmp.ntKjDQB42C /tmp/tmp.rjsIMScLNA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gWERxtc31i +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y5q0fgtNv4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gWERxtc31i ++ cat /tmp/tmp.Y5q0fgtNv4 ++ rm /tmp/tmp.gWERxtc31i /tmp/tmp.Y5q0fgtNv4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l1S53syDNr +++ mktemp ++ local LAST_ERR=/tmp/tmp.unMMYISma6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.l1S53syDNr ++ cat /tmp/tmp.unMMYISma6 ++ rm /tmp/tmp.l1S53syDNr /tmp/tmp.unMMYISma6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5dofKSErT9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.aahkABTH3n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5dofKSErT9 ++ cat /tmp/tmp.aahkABTH3n ++ rm /tmp/tmp.5dofKSErT9 /tmp/tmp.aahkABTH3n ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VMoVLKoINu +++ mktemp ++ local LAST_ERR=/tmp/tmp.VENMqdSoK7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VMoVLKoINu ++ cat /tmp/tmp.VENMqdSoK7 ++ rm /tmp/tmp.VMoVLKoINu /tmp/tmp.VENMqdSoK7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XvRBFwePbA +++ mktemp ++ local LAST_ERR=/tmp/tmp.f2EaGjRJgu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XvRBFwePbA ++ cat /tmp/tmp.f2EaGjRJgu ++ rm /tmp/tmp.XvRBFwePbA /tmp/tmp.f2EaGjRJgu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k3Uh3UrJwb +++ mktemp ++ local LAST_ERR=/tmp/tmp.WrA2rUs3lV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k3Uh3UrJwb ++ cat /tmp/tmp.WrA2rUs3lV ++ rm /tmp/tmp.k3Uh3UrJwb /tmp/tmp.WrA2rUs3lV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NTRsq1ZNCm +++ mktemp ++ local LAST_ERR=/tmp/tmp.bDsdQGD6AU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NTRsq1ZNCm ++ cat /tmp/tmp.bDsdQGD6AU ++ rm /tmp/tmp.NTRsq1ZNCm /tmp/tmp.bDsdQGD6AU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YgofURk58K +++ mktemp ++ local LAST_ERR=/tmp/tmp.gtdIuAydL7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YgofURk58K ++ cat /tmp/tmp.gtdIuAydL7 ++ rm /tmp/tmp.YgofURk58K /tmp/tmp.gtdIuAydL7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nXHDax8P5H +++ mktemp ++ local LAST_ERR=/tmp/tmp.QTZukBxPlD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nXHDax8P5H ++ cat /tmp/tmp.QTZukBxPlD ++ rm /tmp/tmp.nXHDax8P5H /tmp/tmp.QTZukBxPlD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JNXSRbVM3V +++ mktemp ++ local LAST_ERR=/tmp/tmp.9fjyjHxm8z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JNXSRbVM3V ++ cat /tmp/tmp.9fjyjHxm8z ++ rm /tmp/tmp.JNXSRbVM3V /tmp/tmp.9fjyjHxm8z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5gvFAsEb0o +++ mktemp ++ local LAST_ERR=/tmp/tmp.VRzZMhsDGy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5gvFAsEb0o ++ cat /tmp/tmp.VRzZMhsDGy ++ rm /tmp/tmp.5gvFAsEb0o /tmp/tmp.VRzZMhsDGy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QKAKqbB6a6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.hR08QiWvh1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QKAKqbB6a6 ++ cat /tmp/tmp.hR08QiWvh1 ++ rm /tmp/tmp.QKAKqbB6a6 /tmp/tmp.hR08QiWvh1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Kc6dmr5Ddv +++ mktemp ++ local LAST_ERR=/tmp/tmp.QKNSSyI2lk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Kc6dmr5Ddv ++ cat /tmp/tmp.QKNSSyI2lk ++ rm /tmp/tmp.Kc6dmr5Ddv /tmp/tmp.QKNSSyI2lk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3GAguXJ3yq +++ mktemp ++ local LAST_ERR=/tmp/tmp.9jLCN7aOQM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3GAguXJ3yq ++ cat /tmp/tmp.9jLCN7aOQM ++ rm /tmp/tmp.3GAguXJ3yq /tmp/tmp.9jLCN7aOQM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kzHpesPO1Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.AeWEI3ZIzw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kzHpesPO1Y ++ cat /tmp/tmp.AeWEI3ZIzw ++ rm /tmp/tmp.kzHpesPO1Y /tmp/tmp.AeWEI3ZIzw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SYJkK9Y4JE +++ mktemp ++ local LAST_ERR=/tmp/tmp.kDAoCMESIO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SYJkK9Y4JE ++ cat /tmp/tmp.kDAoCMESIO ++ rm /tmp/tmp.SYJkK9Y4JE /tmp/tmp.kDAoCMESIO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ew8DI6wtsy +++ mktemp ++ local LAST_ERR=/tmp/tmp.20jwnmQIDv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ew8DI6wtsy ++ cat /tmp/tmp.20jwnmQIDv ++ rm /tmp/tmp.ew8DI6wtsy /tmp/tmp.20jwnmQIDv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.neZnSqRrGO +++ mktemp ++ local LAST_ERR=/tmp/tmp.kLYXGqqJl6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.neZnSqRrGO ++ cat /tmp/tmp.kLYXGqqJl6 ++ rm /tmp/tmp.neZnSqRrGO /tmp/tmp.kLYXGqqJl6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LifIa6LSMD +++ mktemp ++ local LAST_ERR=/tmp/tmp.Aszvsxmphj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LifIa6LSMD ++ cat /tmp/tmp.Aszvsxmphj ++ rm /tmp/tmp.LifIa6LSMD /tmp/tmp.Aszvsxmphj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0YKYQB62Hc +++ mktemp ++ local LAST_ERR=/tmp/tmp.fgppzbdkvE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0YKYQB62Hc ++ cat /tmp/tmp.fgppzbdkvE ++ rm /tmp/tmp.0YKYQB62Hc /tmp/tmp.fgppzbdkvE ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VdIF1jjah5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.XdeVphAEwV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VdIF1jjah5 ++ cat /tmp/tmp.XdeVphAEwV ++ rm /tmp/tmp.VdIF1jjah5 /tmp/tmp.XdeVphAEwV ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.w0yRbmgrEc ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.arxO3xlbGK +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.w0yRbmgrEc +++++ cat /tmp/tmp.arxO3xlbGK +++++ rm /tmp/tmp.w0yRbmgrEc /tmp/tmp.arxO3xlbGK +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kNqqb5qBG5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Pea1goQA7G ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kNqqb5qBG5 ++ cat /tmp/tmp.Pea1goQA7G ++ rm /tmp/tmp.kNqqb5qBG5 /tmp/tmp.Pea1goQA7G ++ return 0 + [[ 3 == \3 ]] + echo + sleep 15 + compare_mysql_cmd select-3 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local command_id=select-3 + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-3-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SHOW DATABASES;' '-h some-name-haproxy -umonitor -p'\''test-password2'\''' + local 'command=SHOW DATABASES;' + local 'uri=-h some-name-haproxy -umonitor -p'\''test-password2'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mENQeP2wII +++ mktemp ++ local LAST_ERR=/tmp/tmp.2YZOppsALz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mENQeP2wII ++ cat /tmp/tmp.2YZOppsALz ++ rm /tmp/tmp.mENQeP2wII /tmp/tmp.2YZOppsALz ++ return 0 + client_pod=pxc-client-56fd5498cd-2fzfm + wait_pod pxc-client-56fd5498cd-2fzfm + local pod=pxc-client-56fd5498cd-2fzfm + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-2fzfm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-2fzfm condition met waiting for pod/pxc-client-56fd5498cd-2fzfm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.uv48lQ0yvW/select-3.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.uv48lQ0yvW/select-3.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2427/e2e-tests/users-scheduler/compare/select-3.sql /tmp/tmp.uv48lQ0yvW/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + check_generation 1 haproxy some-name + local generation=1 + local container=haproxy + local cluster=some-name + local current_generation ++ kubectl_bin get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VmyH2fxAbi +++ mktemp ++ local LAST_ERR=/tmp/tmp.XZWBbW6ZsH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset some-name-haproxy -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VmyH2fxAbi ++ cat /tmp/tmp.XZWBbW6ZsH ++ rm /tmp/tmp.VmyH2fxAbi /tmp/tmp.XZWBbW6ZsH ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + destroy users-scheduler-10069 + local namespace=users-scheduler-10069 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v level=info + grep -v 'the object has been modified' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + tee /tmp/tmp.uv48lQ0yvW/operator.log ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ + sort -u + grep -v 'get backup status: Job.batch' +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ head -1 ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.xOfB0IXsaw +++ mktemp ++ local LAST_ERR=/tmp/tmp.hgh2IAhH4N ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xOfB0IXsaw ++ cat /tmp/tmp.hgh2IAhH4N ++ rm /tmp/tmp.xOfB0IXsaw /tmp/tmp.hgh2IAhH4N ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-5b5dc68764-h9m2s ++ mktemp + local LAST_OUT=/tmp/tmp.yr8bDdUDW3 ++ mktemp + local LAST_ERR=/tmp/tmp.WcH1bGzFun + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-5b5dc68764-h9m2s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yr8bDdUDW3 + cat /tmp/tmp.WcH1bGzFun + rm /tmp/tmp.yr8bDdUDW3 /tmp/tmp.WcH1bGzFun + return 0 } }, }, { }, }, { }, }, ""), }, { }, - }, - { - }, + }, - "15ced3d8efad2d127cce95c5fd3ae4536a60b3e54ca2bcd68daa4dbea777e402", + "15ced3d8efad2d127cce95c5fd3ae4536a60b3e54ca2bcd68daa4dbea777e402", ... // 16 identical fields ... // 16 identical fields 2026-04-08T14:12:25.200Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "", "enabled": ""} 2026-04-08T14:12:25.200Z INFO setup Manager starting up {"gitCommit": "d4f7121dd0aa338449cb21800d39d94f885e61d4", "gitBranch": "PR-2427-d4f7121d", "buildTime": "2026-04-08T10:03:30Z", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} 2026-04-08T14:12:25.200Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.32.13-gke.1205000"} 2026-04-08T14:12:25.204Z INFO setup Registering Components. 2026-04-08T14:12:25.789Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-04-08T14:12:25.789Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-04-08T14:12:25.789Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-04-08T14:12:25.789Z INFO controller-runtime.metrics Starting metrics server 2026-04-08T14:12:25.789Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-04-08T14:12:25.789Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-04-08T14:12:25.789Z INFO controller-runtime.webhook Starting webhook server 2026-04-08T14:12:25.789Z INFO setup Starting the Cmd. 2026-04-08T14:12:25.789Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-04-08T14:12:25.890Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-08T14:12:25.919Z DEBUG events percona-xtradb-cluster-operator-5b5dc68764-h9m2s_a8c835d2-7754-48dc-94c4-65a11062b9e1 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"2492b77b-1621-4275-a93f-8c72386d83ae","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1775657545912863009"}, "reason": "LeaderElection"} 2026-04-08T14:12:25.919Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-04-08T14:12:25.919Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-04-08T14:12:25.919Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-08T14:12:25.920Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-04-08T14:12:25.920Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-04-08T14:12:26.021Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-04-08T14:12:26.021Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-04-08T14:12:26.021Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-04-08T14:12:26.021Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-04-08T14:12:26.021Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-04-08T14:12:26.021Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-04-08T14:12:40.350Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "version": "1.20.0"} 2026-04-08T14:12:40.491Z INFO User secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "secrets": "my-cluster-secrets"} 2026-04-08T14:12:40.711Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-08T14:12:40.831Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-08T14:12:40.864Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-08T14:12:40.916Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:12:40.965Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:12:41.015Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:12:41.099Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "a35a5647-bfe8-4af2-aba7-c840951614e7", "object": "some-name-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:12:41.917Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "d13d82b1-0f08-438a-98d2-0c6770ed5a9e", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-08T14:12:41.936Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "d13d82b1-0f08-438a-98d2-0c6770ed5a9e", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-08T14:13:21.792Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "7fcd20ef-d159-4238-8e0a-58cfdfb16a9e", "error": "reconcile users secret: create Users secret: secrets \"my-cluster-secrets\" is forbidden: unable to create new content in namespace users-scheduler-31309 because it is being terminated", "errorVerbose": "create Users secret: secrets \"my-cluster-secrets\" is forbidden: unable to create new content in namespace users-scheduler-31309 because it is being terminated\nreconcile users secret\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:330\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:13:29.437Z ERROR Update status {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-31309"}, "namespace": "users-scheduler-31309", "name": "some-name", "reconcileID": "3f30d65c-85d0-46d4-a6d9-a94bbd06b4c4", "error": "PerconaXtraDBCluster.pxc.percona.com \"some-name\" not found"} 2026-04-08T14:13:29.466Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "989154e9-6c0b-4007-af7a-6dc4a7418513", "version": "1.20.0"} 2026-04-08T14:13:29.619Z INFO User secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "989154e9-6c0b-4007-af7a-6dc4a7418513", "secrets": "my-cluster-secrets"} 2026-04-08T14:13:29.637Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "989154e9-6c0b-4007-af7a-6dc4a7418513", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-08T14:13:29.652Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "989154e9-6c0b-4007-af7a-6dc4a7418513", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-08T14:13:30.204Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "989154e9-6c0b-4007-af7a-6dc4a7418513", "error": "failed to deploy proxysql: updatePod for proxysql: reconcile config: reconcile autotune config: create or update configmap: configmaps \"auto-some-name-pxc\" already exists", "errorVerbose": "configmaps \"auto-some-name-pxc\" already exists\ncreate or update configmap\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileAutotuneConfigMap\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:92\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:25\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:48\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:542\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile autotune config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:27\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:48\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:542\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:50\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:542\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nupdatePod for proxysql\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:543\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:560\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nfailed to deploy proxysql\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:561\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:367\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:13:30.310Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f402a8a1-a116-4edd-9c45-2261838e6a7e", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-08T14:13:30.348Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f402a8a1-a116-4edd-9c45-2261838e6a7e", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-08T14:13:30.439Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f402a8a1-a116-4edd-9c45-2261838e6a7e", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:13:30.486Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f402a8a1-a116-4edd-9c45-2261838e6a7e", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:13:30.532Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f402a8a1-a116-4edd-9c45-2261838e6a7e", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:13:30.650Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f402a8a1-a116-4edd-9c45-2261838e6a7e", "object": "some-name-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:13:31.307Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "3138d7cf-dc91-4526-9026-b7f526ac0175", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-08T14:13:31.324Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "3138d7cf-dc91-4526-9026-b7f526ac0175", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-08T14:14:58.001Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6", "user": "operator"} 2026-04-08T14:14:58.036Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6", "user": "monitor"} 2026-04-08T14:14:58.103Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6"} 2026-04-08T14:14:58.139Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6"} 2026-04-08T14:14:58.171Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6", "user": "xtrabackup"} 2026-04-08T14:14:58.221Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6"} 2026-04-08T14:14:58.258Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6", "user": "replication"} 2026-04-08T14:14:58.267Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "367d21e1-6dbf-41e1-87fb-66d995cf2bb6", "err": "get primary pxc pod: not found"} 2026-04-08T14:15:03.381Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "3e377c41-c2a1-42e5-adbf-5db62dd5c689", "err": "get primary pxc pod: not found"} 2026-04-08T14:15:08.477Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0027f4be-d130-4244-a4f3-f427c6a1dcce", "err": "get primary pxc pod: not found"} 2026-04-08T14:17:45.514Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "725fc071-88d5-4568-9ee1-8b052146125e", "user": "root"} 2026-04-08T14:17:45.583Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "725fc071-88d5-4568-9ee1-8b052146125e", "new version": "8.4.7-7.1"} 2026-04-08T14:17:48.124Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "725fc071-88d5-4568-9ee1-8b052146125e"} 2026-04-08T14:17:54.031Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "5d05e528-8011-40a4-8f38-f4852d56c23f"} 2026-04-08T14:17:59.803Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "26f252dc-e6df-4baa-a780-e9773fb85624"} 2026-04-08T14:18:05.214Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "6487fe00-6e19-46e1-bb7e-7d1f39ee25e2"} 2026-04-08T14:18:10.600Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "4e2598ae-b1ad-4f2b-bf79-e5b89a284704"} 2026-04-08T14:18:16.029Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d97dcbbd-f1c2-405e-9ca1-8d593ed2556e"} 2026-04-08T14:18:21.210Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "08c1c506-e471-4672-adbf-7202e93d74da"} 2026-04-08T14:18:26.543Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "ead087d3-a621-46d7-81a0-89b7dad2dac7"} 2026-04-08T14:18:31.609Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "8799f228-32d8-4716-bcc2-2976fa8fe1fc"} 2026-04-08T14:18:37.203Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "4fc28f43-dc56-47fe-a3e0-87f1f66f4239"} 2026-04-08T14:18:42.522Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "20f49960-5b06-4b13-8ba7-b01845c5e1af"} 2026-04-08T14:18:47.797Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "469262aa-f9aa-4947-b5db-25adf1fc934e"} 2026-04-08T14:18:52.934Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "114a6eb4-df23-4ac3-b96e-740178091ef6"} 2026-04-08T14:18:58.595Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b0031e10-b301-4ce3-aafb-2d31598afb8c"} 2026-04-08T14:19:03.523Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "78bda2a0-cf86-4b7e-805c-10e1474b22c8"} 2026-04-08T14:19:08.846Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "13a01d1e-6571-4d55-9eb3-d3edeac1fdf5"} 2026-04-08T14:19:14.895Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "03c16881-e05a-4cad-b784-682d90717bd4"} 2026-04-08T14:19:19.810Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "4a9472a8-bf2b-4418-96f5-55d1dae99596"} 2026-04-08T14:19:24.943Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b945f7a8-000d-45e0-9e7f-4fb9efb8f588"} 2026-04-08T14:19:30.816Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "33bb33ad-1c74-468a-9fa8-1092495835cf"} 2026-04-08T14:19:33.715Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a45bb340-7f24-45d6-8110-eb1f0cce6d31", "user": "root"} 2026-04-08T14:19:33.753Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a45bb340-7f24-45d6-8110-eb1f0cce6d31", "user": "root"} 2026-04-08T14:19:33.787Z INFO MySQL init secret created {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a45bb340-7f24-45d6-8110-eb1f0cce6d31", "secret": "some-name-mysql-init", "user": "root"} 2026-04-08T14:19:38.718Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "afdaf167-c642-42e4-b146-fd49d25506d2"} 2026-04-08T14:19:38.894Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a45bb340-7f24-45d6-8110-eb1f0cce6d31"} 2026-04-08T14:19:39.361Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a45bb340-7f24-45d6-8110-eb1f0cce6d31", "user": "root"} 2026-04-08T14:19:39.392Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a45bb340-7f24-45d6-8110-eb1f0cce6d31", "user": "root"} 2026-04-08T14:19:41.741Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a45bb340-7f24-45d6-8110-eb1f0cce6d31"} 2026-04-08T14:19:47.239Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "32e20894-28a4-4e94-8f1f-30fe6915c0bb"} 2026-04-08T14:19:52.419Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d2d2d193-b12c-4b44-bb6e-8e79e7ab680f"} 2026-04-08T14:19:57.724Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "5f5d4376-7961-4421-929d-23e00415507d"} 2026-04-08T14:20:03.199Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "17709088-ca80-4201-a740-9fe20d959af9"} 2026-04-08T14:20:08.306Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "93bff802-389a-4fd7-968d-ef933bc450f5"} 2026-04-08T14:20:13.715Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "3e4bae0d-5808-439d-a7a0-a6930ef8e195"} 2026-04-08T14:20:18.933Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "915e78be-5e59-47a8-ac93-bdc45c284d55"} 2026-04-08T14:20:21.018Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "be47c7d4-6f97-47b0-92af-b2e4640fa18c", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:20:21.109Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "be47c7d4-6f97-47b0-92af-b2e4640fa18c", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:20:24.663Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "be47c7d4-6f97-47b0-92af-b2e4640fa18c", "error": "exec syncusers: failed to execute command in pod: pod some-name-proxysql-2 does not have a host assigned / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: pod some-name-proxysql-2 does not have a host assigned / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:20:53.804Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "dea4ecdd-8553-4f41-ba54-5dfb58d536c8", "user": "proxyadmin"} 2026-04-08T14:20:53.804Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "dea4ecdd-8553-4f41-ba54-5dfb58d536c8", "user": "proxyadmin"} 2026-04-08T14:20:53.860Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "dea4ecdd-8553-4f41-ba54-5dfb58d536c8", "user": "proxyadmin"} 2026-04-08T14:20:53.878Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "dea4ecdd-8553-4f41-ba54-5dfb58d536c8", "user": "proxyadmin"} 2026-04-08T14:20:53.878Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "dea4ecdd-8553-4f41-ba54-5dfb58d536c8", "last-applied-secret": "4bcdb6fea9c0299d4331d4c1c4912d8d200ecda89983a7799bdf3b49d6c91d98"} 2026-04-08T14:20:53.882Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "dea4ecdd-8553-4f41-ba54-5dfb58d536c8", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:20:54.709Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c344e804-623a-4176-bd8d-c2a354300e76", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\n", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\nERROR (line:515) : The cluster (with writer hostgroup:11) has not been configured in ProxySQL\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:22:01.396Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e546b154-92c2-46fc-b85e-6637958152f7"} 2026-04-08T14:22:06.884Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "ff98f5a3-e32a-4f1f-bd39-b1876a556250"} 2026-04-08T14:22:11.776Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b3f23286-1740-4719-bb37-cfa21d1f6b82"} 2026-04-08T14:22:17.286Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "39d49b3c-5018-4460-ac8a-b63a5cf5f800"} 2026-04-08T14:22:22.421Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e54bb874-30e2-4f6a-ae20-ed6ada547fbb"} 2026-04-08T14:22:25.105Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "6b672a77-2d5b-42ed-b49e-0d6613aa6bbb", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:22:27.529Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c735e498-2144-4501-b6cc-9712b824c4d3", "error": "exec syncusers: failed to execute command in pod: pods \"some-name-proxysql-2\" not found / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: pods \"some-name-proxysql-2\" not found / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:22:33.066Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1", "user": "xtrabackup"} 2026-04-08T14:22:33.090Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1", "user": "xtrabackup"} 2026-04-08T14:22:33.114Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-08T14:22:33.130Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1", "user": "xtrabackup"} 2026-04-08T14:22:33.149Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1", "user": "xtrabackup"} 2026-04-08T14:22:33.152Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1", "last-applied-secret": "dbe414e5bbd906192d94a07211d598cbb4276befe30312deb0732504ffc62d98"} 2026-04-08T14:22:33.156Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:22:36.320Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "d7271cc6-0e56-4417-bf19-83e1724cc9a1"} 2026-04-08T14:22:38.732Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0c8114ad-ec55-4821-b37d-9d0e7130c225", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp: lookup some-name-pxc-2.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:22:41.001Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0c8114ad-ec55-4821-b37d-9d0e7130c225"} 2026-04-08T14:23:53.005Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e473ce96-3bd1-4b3e-8744-5e7ccc1eef29", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp 10.148.74.63:33062: connect: connection refused"} 2026-04-08T14:24:46.098Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "458aa46e-5a83-449f-8ffd-e3fe7ac8fb8a", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:24:51.357Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0c2582d7-69ad-4a22-b7d9-8fd7229d48ba", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:24:56.507Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "3e890c15-d603-4b3e-b2d6-8cdff4d3c046", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:01.675Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "814a7a77-bdce-41e4-9869-5b8881252679", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:06.864Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a10dbc16-fbc3-489c-8d70-77b5fa028eea", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:12.028Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "8231d1d7-dc6e-4561-9ed2-406b212701f5", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:17.247Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "aa3b5eb4-2c9e-467c-ac72-398cac93d150", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:22.445Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a9e7fa28-f3b3-4451-ab81-280789c1ee33", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:27.567Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "9d396e9e-b610-4029-8d9b-e2a6669e85cf", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:32.696Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "259f5f1b-4d9d-4915-baab-128ef39b77c3", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:25:42.412Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "5fe68ca8-9714-42d7-8911-238ae7cafbc5"} 2026-04-08T14:25:46.590Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a8fd85d0-a8cb-436b-ac3a-7ca4b1b54f59"} 2026-04-08T14:25:51.797Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "fa0c9b00-9c05-4a45-a5de-860275795578"} 2026-04-08T14:25:57.271Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "58f2b3bf-427c-4004-8534-cab1e94e583c"} 2026-04-08T14:26:01.061Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e1727827-9f72-483f-bff7-974f1ca7c9cf", "user": "monitor"} 2026-04-08T14:26:01.085Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e1727827-9f72-483f-bff7-974f1ca7c9cf", "user": "monitor"} 2026-04-08T14:26:01.113Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e1727827-9f72-483f-bff7-974f1ca7c9cf", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-08T14:26:01.187Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e1727827-9f72-483f-bff7-974f1ca7c9cf", "user": "monitor"} 2026-04-08T14:26:01.209Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e1727827-9f72-483f-bff7-974f1ca7c9cf", "user": "monitor"} 2026-04-08T14:26:01.301Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e1727827-9f72-483f-bff7-974f1ca7c9cf", "last-applied-secret": "9c37bdb7cce228df3c6dffa335c6ebc05b2269431c8716e28484b4d3290bfdd4"} 2026-04-08T14:26:01.305Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e1727827-9f72-483f-bff7-974f1ca7c9cf", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:26:02.187Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "68592108-b139-4a90-a166-7dfbb47e51cf", "error": "exec syncusers: failed to execute command in pod: pods \"some-name-proxysql-1\" not found / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: pods \"some-name-proxysql-1\" not found / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:26:04.152Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "afe0afc0-7751-487d-a749-c82cd3317348", "user": "monitor"} 2026-04-08T14:26:04.455Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "afe0afc0-7751-487d-a749-c82cd3317348", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:26:04.507Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "afe0afc0-7751-487d-a749-c82cd3317348", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:26:06.894Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "afe0afc0-7751-487d-a749-c82cd3317348", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:26:53.123Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "386370a4-b5b9-4365-bd94-f119993d39aa", "user": "monitor"} 2026-04-08T14:26:56.958Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "386370a4-b5b9-4365-bd94-f119993d39aa"} 2026-04-08T14:26:58.974Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "db34984c-6040-471e-b1e8-2630019206d3", "user": "monitor"} 2026-04-08T14:27:02.617Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "db34984c-6040-471e-b1e8-2630019206d3"} 2026-04-08T14:27:04.521Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b5c2dbd6-ffe8-4562-8f9f-5bc3ed7c6d18", "user": "monitor"} 2026-04-08T14:27:05.399Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b5c2dbd6-ffe8-4562-8f9f-5bc3ed7c6d18", "user": "monitor"} 2026-04-08T14:27:05.405Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b5c2dbd6-ffe8-4562-8f9f-5bc3ed7c6d18", "last-applied-secret": "9c37bdb7cce228df3c6dffa335c6ebc05b2269431c8716e28484b4d3290bfdd4"} 2026-04-08T14:27:08.959Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b5c2dbd6-ffe8-4562-8f9f-5bc3ed7c6d18"} 2026-04-08T14:27:14.211Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "7af5dd95-11da-4e91-81c3-e78fb256fe7d"} 2026-04-08T14:27:19.820Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f887cb5d-0a0b-4522-88b9-a443bcb6ad21"} 2026-04-08T14:27:25.040Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "ba04c9a1-774a-4f17-ba9a-3a8811c8dfb5"} 2026-04-08T14:27:29.941Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "ac15d21a-7d69-46d2-abef-e63e08b9ebef"} 2026-04-08T14:27:35.047Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "eb9874bc-8da3-423d-8413-d79638893f07"} 2026-04-08T14:27:40.551Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "12c0ebf5-0158-4e48-8ee1-6cb3d2d21146"} 2026-04-08T14:27:44.993Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "64eaf04f-f832-4911-9e97-89de7d5e5a61", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:27:46.472Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "7d0505c1-c040-4eff-b281-c05c2142a7ac", "error": "exec syncusers: failed to execute command in pod: pods \"some-name-proxysql-2\" not found / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: pods \"some-name-proxysql-2\" not found / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:27:49.208Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "47cf9aea-7dc5-4ec7-a16e-d0e5ba866a8f"} 2026-04-08T14:27:59.730Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "eb0db7b2-083c-436e-9799-10a3134f4deb"} 2026-04-08T14:28:05.311Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "839fd823-d35e-4d7d-bc5e-697ccff7c954"} 2026-04-08T14:28:10.409Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "6000aa4e-7c4d-4d9d-b4fe-a465d3b32f1e"} 2026-04-08T14:28:12.008Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "user": "operator"} 2026-04-08T14:28:12.032Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "user": "operator"} 2026-04-08T14:28:12.053Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-08T14:28:12.073Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "user": "operator"} 2026-04-08T14:28:12.093Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "user": "operator"} 2026-04-08T14:28:12.108Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "last-applied-secret": "7032c6eb8d0a7de1f96805dd0db3fca084ad05355835b06df63277bb9bf655c7"} 2026-04-08T14:28:12.112Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:28:17.599Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "659f4e08-af15-4b1c-b4da-50e6f429b135", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:28:55.023Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "9bc8a06d-aa2d-45ab-a8f2-b9258843c838"} 2026-04-08T14:29:00.141Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "7bed9bd0-3d59-4943-a6d6-c523fa0248e9"} 2026-04-08T14:29:05.026Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a1f7f5c8-4d89-4088-ab24-ab040a74c711"} 2026-04-08T14:29:10.224Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f70ea84d-f78a-44a3-a892-984d986dcf96"} 2026-04-08T14:29:15.826Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "55f8017e-7a8b-44eb-bb1d-a448dbe9ad46"} 2026-04-08T14:29:17.816Z INFO Created user secrets {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "secrets": "my-cluster-secrets-2"} 2026-04-08T14:29:17.821Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "root"} 2026-04-08T14:29:17.860Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "root"} 2026-04-08T14:29:17.890Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "secret": "some-name-mysql-init", "user": "root"} 2026-04-08T14:29:20.805Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60"} 2026-04-08T14:29:20.827Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "root"} 2026-04-08T14:29:20.859Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "root"} 2026-04-08T14:29:20.863Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "operator"} 2026-04-08T14:29:20.884Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "operator"} 2026-04-08T14:29:20.905Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-08T14:29:20.930Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "operator"} 2026-04-08T14:29:20.947Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "operator"} 2026-04-08T14:29:20.954Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "monitor"} 2026-04-08T14:29:20.974Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "monitor"} 2026-04-08T14:29:20.991Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-08T14:29:21.023Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "monitor"} 2026-04-08T14:29:21.045Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "monitor"} 2026-04-08T14:29:21.342Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "xtrabackup"} 2026-04-08T14:29:21.362Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "xtrabackup"} 2026-04-08T14:29:21.381Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-08T14:29:21.402Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "xtrabackup"} 2026-04-08T14:29:21.421Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "xtrabackup"} 2026-04-08T14:29:21.425Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "replication"} 2026-04-08T14:29:21.445Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "replication"} 2026-04-08T14:29:21.464Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "secret": "some-name-mysql-init", "user": "replication"} 2026-04-08T14:29:21.485Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "replication"} 2026-04-08T14:29:21.504Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "replication"} 2026-04-08T14:29:21.504Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "proxyadmin"} 2026-04-08T14:29:21.541Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "proxyadmin"} 2026-04-08T14:29:21.563Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "user": "proxyadmin"} 2026-04-08T14:29:21.563Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "last-applied-secret": "cf65cfece5bc103bdae462b53efda8815b7150602e94729558a5c5d9ce908d04"} 2026-04-08T14:29:21.563Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "last-applied-secret": "cf65cfece5bc103bdae462b53efda8815b7150602e94729558a5c5d9ce908d04"} 2026-04-08T14:29:21.566Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:29:21.680Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:29:25.550Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c834c1cb-4ef7-4350-811c-19ab3f989e60", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.108' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:30:37.114Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0f1bdd3c-ef62-42f6-8e5b-e0a591d25dac", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp 10.148.74.66:33062: connect: connection refused"} 2026-04-08T14:31:34.835Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "9583d3c2-0708-4a58-b179-a22be92d379a", "err": "failed to connect to pod some-name-pxc-0: dial tcp 10.148.72.106:33062: connect: connection refused"} 2026-04-08T14:31:40.002Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c82fe031-d63e-4d49-9756-7fe7cad2e8d2", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:31:45.105Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c4bd2451-2e9a-45b2-9b3f-48c925167038", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:31:50.210Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f4011c86-21a0-4e09-aa9e-f1db4950f121", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:31:55.319Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "21f832de-69bb-46c0-993d-7ad6a2d86e68", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:32:00.441Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "7fc89952-9af2-4963-a2fe-c20a00c179ef", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:32:05.568Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c62767ec-86c8-41f1-868f-391fc08f169f", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:32:10.705Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "bf62a295-7df1-492b-8250-a7ca6aebb251", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:32:15.810Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "f962572d-30d5-4904-8dbd-c2f5b258ea6a", "primary name": "some-name-pxc-0.some-name-pxc.users-scheduler-10069.svc.cluster.local"} 2026-04-08T14:32:21.743Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "67b9e03e-dfe9-40af-acb4-5b638b159a55", "user": "monitor"} 2026-04-08T14:32:22.229Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "67b9e03e-dfe9-40af-acb4-5b638b159a55", "user": "monitor"} 2026-04-08T14:32:22.237Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "67b9e03e-dfe9-40af-acb4-5b638b159a55", "last-applied-secret": "cf65cfece5bc103bdae462b53efda8815b7150602e94729558a5c5d9ce908d04"} 2026-04-08T14:32:26.622Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "67b9e03e-dfe9-40af-acb4-5b638b159a55"} 2026-04-08T14:32:26.872Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "user": "operator"} 2026-04-08T14:32:26.894Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "user": "operator"} 2026-04-08T14:32:26.915Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-08T14:32:26.937Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "user": "operator"} 2026-04-08T14:32:26.958Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "user": "operator"} 2026-04-08T14:32:26.969Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "last-applied-secret": "15ced3d8efad2d127cce95c5fd3ae4536a60b3e54ca2bcd68daa4dbea777e402"} 2026-04-08T14:32:26.973Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:32:32.247Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b31c791f-73ed-4336-a572-e3cac90c1de3", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR 1045 (28000): Access denied for user 'operator'@'some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-' (using password: YES)\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:33:06.506Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "30869063-2c03-424a-999d-73b561061472"} 2026-04-08T14:33:11.477Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a4068a22-9492-4ed4-a19c-6c3f5e3a58bf"} 2026-04-08T14:33:16.799Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "33006f37-6033-4076-b725-06a5f2ce1b9a"} 2026-04-08T14:33:21.990Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "bdfc0f6a-c680-42fc-8b89-165da31f21ed"} 2026-04-08T14:33:26.995Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "35a0b196-8e3d-4f18-a19e-18b69e3bf1e4"} 2026-04-08T14:33:32.489Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "834c84a4-5e2c-4f80-a1da-f9758f9ed1ac"} 2026-04-08T14:33:37.884Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "1a6c7c20-c9a8-49b1-a9fa-924d8c22499f"} 2026-04-08T14:33:43.755Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "098b2a1c-6c2e-4dc3-aed6-e75d7cb28ec1"} 2026-04-08T14:33:48.310Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b23fce94-5d69-4145-bdc5-c6b57490d2d5"} 2026-04-08T14:33:53.514Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "83ba8d97-3754-488f-a65a-7523d27527c9"} 2026-04-08T14:33:59.005Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "fab54e41-0039-4129-8e9f-afcdd3683f16"} 2026-04-08T14:34:04.091Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "8412f0f9-c311-4b85-8499-12e93af1af73"} 2026-04-08T14:34:09.412Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "023fa22e-47ae-4cce-8185-0751f06c75f9"} 2026-04-08T14:34:14.919Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "ad1570b2-a5bb-4aab-a471-caeaf71c46fb"} 2026-04-08T14:34:20.410Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "8f284028-0876-465f-829c-3abe88b93ee6"} 2026-04-08T14:34:25.672Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "01b46d2a-c2df-4e9b-b6c3-d36298598ed6"} 2026-04-08T14:34:30.898Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "98090dcf-72b8-4dee-8c84-d0dd900f4f51"} 2026-04-08T14:34:36.298Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "77e6cb2f-6f4a-41c0-870c-5a55f684a222"} 2026-04-08T14:34:41.562Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0d9c5ca8-f69e-48e9-b22c-d1c6158e13a9"} 2026-04-08T14:34:46.772Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "370a5af1-58e4-4a13-9e3d-5f39b87d2444"} 2026-04-08T14:34:51.990Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "b45f4a39-bf0a-44ca-8b8a-2389edbd87f8"} 2026-04-08T14:34:57.316Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "4f0dd231-a552-44ee-aac7-2bdfacfa8cf9"} 2026-04-08T14:35:02.661Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "dfa41eb6-1823-4822-990c-3b345db42d40"} 2026-04-08T14:35:04.251Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "root"} 2026-04-08T14:35:04.293Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "root"} 2026-04-08T14:35:04.313Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "secret": "some-name-mysql-init", "user": "root"} 2026-04-08T14:35:07.495Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0"} 2026-04-08T14:35:07.515Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "root"} 2026-04-08T14:35:07.548Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "root"} 2026-04-08T14:35:07.555Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "monitor"} 2026-04-08T14:35:07.577Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "monitor"} 2026-04-08T14:35:07.600Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-08T14:35:07.633Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "monitor"} 2026-04-08T14:35:07.658Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "monitor"} 2026-04-08T14:35:07.947Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "xtrabackup"} 2026-04-08T14:35:07.969Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "xtrabackup"} 2026-04-08T14:35:07.993Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-08T14:35:08.015Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "xtrabackup"} 2026-04-08T14:35:08.034Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "xtrabackup"} 2026-04-08T14:35:08.038Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "proxyadmin"} 2026-04-08T14:35:08.073Z INFO Proxy user updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "proxyadmin"} 2026-04-08T14:35:08.094Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "user": "proxyadmin"} 2026-04-08T14:35:08.094Z INFO Proxy pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "last-applied-secret": "dc69a41ea25f70c2e5f67654b729881ec47dd97467f549c879c4d6149bfd0e75"} 2026-04-08T14:35:08.094Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "last-applied-secret": "dc69a41ea25f70c2e5f67654b729881ec47dd97467f549c879c4d6149bfd0e75"} 2026-04-08T14:35:08.097Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:35:08.179Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:35:11.363Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "a86e2552-5a19-4343-9eae-4a826a610fc0", "error": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n", "errorVerbose": "exec syncusers: failed to execute command in pod: command terminated with exit code 1 / / ERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\nERROR 1045 (28000): ProxySQL Error: Access denied for user 'proxyadmin'@'10.148.72.111' (using password: YES)\nERROR (line:586) : ProxySQL connection check failed. \n-- Could not connect to ProxySQL at some-name-proxysql-0.some-name-proxysql-unready.users-scheduler-10069.svc.cluster.local:6032 \n-- Please check the ProxySQL connection parameters and status.\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:975\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-08T14:36:08.645Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c159469b-b1fa-43ea-a598-0fdda836a648", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-08T14:36:08.696Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c159469b-b1fa-43ea-a598-0fdda836a648", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-08T14:36:08.762Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c159469b-b1fa-43ea-a598-0fdda836a648", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:36:08.862Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "c159469b-b1fa-43ea-a598-0fdda836a648", "object": "some-name-haproxy-replicas", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-08T14:36:09.415Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "9642ba3b-9968-48c1-a30d-9718cd0d3552", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-08T14:36:18.021Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "5d8dd465-7674-45b0-809b-6de3b5224bed", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 34.118.231.18:3306: connect: connection refused"} 2026-04-08T14:36:26.245Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "1f302fec-bdcf-4fbe-bbd5-c84ddc0940ce", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 34.118.231.18:3306: connect: connection refused"} 2026-04-08T14:36:34.518Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "43eb29e8-2a08-49b8-ab52-0bff2dfa144e", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 34.118.231.18:3306: connect: connection refused"} 2026-04-08T14:37:11.252Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "fddb1f03-5b55-48bf-9ee8-dc77532dbe8f", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:37:11.544Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "ab192bb5-894a-45e7-b523-a8aec2f2028a", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:37:17.140Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "8d0689e1-2ee6-4c30-9ae3-718e1ea4b650", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:38:07.506Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "root"} 2026-04-08T14:38:07.546Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "root"} 2026-04-08T14:38:07.583Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "secret": "some-name-mysql-init", "user": "root"} 2026-04-08T14:38:07.607Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "root"} 2026-04-08T14:38:07.636Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "root"} 2026-04-08T14:38:07.639Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "operator"} 2026-04-08T14:38:07.660Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "operator"} 2026-04-08T14:38:07.692Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "secret": "some-name-mysql-init", "user": "operator"} 2026-04-08T14:38:07.711Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "operator"} 2026-04-08T14:38:07.730Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "operator"} 2026-04-08T14:38:07.733Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "monitor"} 2026-04-08T14:38:07.756Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "monitor"} 2026-04-08T14:38:07.784Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-08T14:38:07.822Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "monitor"} 2026-04-08T14:38:08.133Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "xtrabackup"} 2026-04-08T14:38:08.154Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "xtrabackup"} 2026-04-08T14:38:08.176Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "secret": "some-name-mysql-init", "user": "xtrabackup"} 2026-04-08T14:38:08.208Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "xtrabackup"} 2026-04-08T14:38:08.226Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "xtrabackup"} 2026-04-08T14:38:08.230Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "replication"} 2026-04-08T14:38:08.249Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "replication"} 2026-04-08T14:38:08.275Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "secret": "some-name-mysql-init", "user": "replication"} 2026-04-08T14:38:08.302Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "replication"} 2026-04-08T14:38:08.320Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "user": "replication"} 2026-04-08T14:38:08.320Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "last-applied-secret": "7032c6eb8d0a7de1f96805dd0db3fca084ad05355835b06df63277bb9bf655c7"} 2026-04-08T14:38:08.322Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e382974b-c4c8-48c7-9c5b-15d4f76e16b3", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-08T14:40:07.380Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "3d0fbdcb-e3c5-4356-82ef-c538d7ae6d53", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:40:08.351Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "aa595e57-e177-436c-afaf-07f548599ab5", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:40:29.942Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "fdd40ccd-73ba-4db0-b80b-c4c3d97163a1", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.users-scheduler-10069 on 34.118.224.10:53: no such host"} 2026-04-08T14:40:57.128Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0d9ded12-05b9-4ee3-9755-13016e046dcb", "user": "monitor"} 2026-04-08T14:40:57.150Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0d9ded12-05b9-4ee3-9755-13016e046dcb", "user": "monitor"} 2026-04-08T14:40:57.174Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0d9ded12-05b9-4ee3-9755-13016e046dcb", "secret": "some-name-mysql-init", "user": "monitor"} 2026-04-08T14:40:57.220Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "0d9ded12-05b9-4ee3-9755-13016e046dcb", "user": "monitor"} 2026-04-08T14:41:02.918Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "ea2e2bc4-1787-4e6b-9d98-f2e7ba446a0b", "user": "monitor"} 2026-04-08T14:41:08.602Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "75016607-1f10-4b73-b33a-0d13bb9a0a20", "user": "monitor"} 2026-04-08T14:41:14.039Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "e8ca049a-8096-4366-83ce-e161cceec3b2", "user": "monitor"} 2026-04-08T14:41:19.711Z INFO Password updated but old one not discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"users-scheduler-10069"}, "namespace": "users-scheduler-10069", "name": "some-name", "reconcileID": "8398fdf7-dde5-482d-a58f-a93659c93e65", "user": "monitor"} ... // 22 identical fields ... // 2 identical fields ... // 3 identical fields ... // 3 identical fields ... // 3 identical fields - "4bcdb6fea9c0299d4331d4c1c4912d8d200ecda89983a7799bdf3b49d6c91d98", ... // 4 identical fields ... // 5 identical fields ... // 6 identical fields - "7032c6eb8d0a7de1f96805dd0db3fca084ad05355835b06df63277bb9bf655c7", + "7032c6eb8d0a7de1f96805dd0db3fca084ad05355835b06df63277bb9bf655c7", ... // 7 identical fields ... // 8 identical fields - "9c37bdb7cce228df3c6dffa335c6ebc05b2269431c8716e28484b4d3290bfdd4", + "9c37bdb7cce228df3c6dffa335c6ebc05b2269431c8716e28484b4d3290bfdd4", ... // 9 identical fields ... // 9 identical fields AccessModes: nil, ActiveDeadlineSeconds: nil, Affinity: nil, Annotations: map[string]string{ - Annotations: map[string]string{ + Annotations: map[string]string{ - APIVersion: "apps/v1", - APIVersion: "apps/v1", Args: {"mysqld"}, Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...}, AutomountServiceAccountToken: nil, + AvailableReplicas: 0, - AvailableReplicas: 1, - AvailableReplicas: 2, - AvailableReplicas: 3, AWSElasticBlockStore: nil, AzureFile: nil, Capacity: nil, - "cf65cfece5bc103bdae462b53efda8815b7150602e94729558a5c5d9ce908d04", + "cf65cfece5bc103bdae462b53efda8815b7150602e94729558a5c5d9ce908d04", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, ConfigMap: &v1.ConfigMapVolumeSource{ ContainerPort: 3306, ContainerPort: 33060, ContainerPort: 33062, ContainerPort: 4444, ContainerPort: 4567, ContainerPort: 4568, ContainerPort: 6032, ContainerPort: 6070, Containers: []v1.Container{ + CreationTimestamp: v1.Time{}, - CreationTimestamp: v1.Time{Time: s"2026-04-08 14:13:30 +0000 UTC"}, + CurrentReplicas: 0, - CurrentReplicas: 1, - CurrentReplicas: 2, - CurrentReplicas: 3, + CurrentRevision: "", - CurrentRevision: "some-name-proxysql-55c8f97957", - CurrentRevision: "some-name-proxysql-5857b64f7c", - CurrentRevision: "some-name-proxysql-5d7957976", - CurrentRevision: "some-name-proxysql-6876776859", - CurrentRevision: "some-name-proxysql-849557b8c6", - CurrentRevision: "some-name-proxysql-99559dd8b", - CurrentRevision: "some-name-pxc-7574cf9956", - CurrentRevision: "some-name-pxc-7865f4f8d", - CurrentRevision: "some-name-pxc-7c489f5c8", - CurrentRevision: "some-name-pxc-8565bdfdd", DataSource: nil, DataSourceRef: nil, - "dbe414e5bbd906192d94a07211d598cbb4276befe30312deb0732504ffc62d98", - "dc69a41ea25f70c2e5f67654b729881ec47dd97467f549c879c4d6149bfd0e75", + "dc69a41ea25f70c2e5f67654b729881ec47dd97467f549c879c4d6149bfd0e75", - DefaultMode: &420, - DefaultMode: &420, + DefaultMode: nil, + DefaultMode: nil, DeletionGracePeriodSeconds: nil, DeletionGracePeriodSeconds: nil, DeletionTimestamp: nil, + DeprecatedServiceAccount: "", - DeprecatedServiceAccount: "default", + DNSPolicy: "", - DNSPolicy: "ClusterFirst", EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-env-vars-proxysql"}, Optional: &true}}}, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "some-name-env-vars-pxc"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "some-name-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "proxyadmin"}}}, ...}, Env: {{Name: "PXC_SERVICE", Value: "some-name-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-some-name"}, Key: "xtrabackup"}}}, ...}, EphemeralContainers: nil, FailureThreshold: 3, FC: nil, - FieldsType: "FieldsV1", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., Finalizers: nil, Finalizers: nil, + Generation: 0, - Generation: 1, - Generation: 10, - Generation: 2, - Generation: 3, - Generation: 4, - Generation: 5, - Generation: 6, - Generation: 7, - Generation: 8, - Generation: 9, github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile.func1 github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 GitRepo: nil, /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:259 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:484 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:834 HostAliases: nil, HostIP: "", HostPort: 0, ImagePullPolicy: "Always", InitContainers: []v1.Container{ InitialDelaySeconds: 300, ISCSI: nil, Items: nil, Items: nil, "kubectl.kubernetes.io/default-container": "proxysql", "kubectl.kubernetes.io/default-container": "pxc", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Labels: nil, + "last-applied-secret": "4bcdb6fea9c0299d4331d4c1c4912d8d200ecda89983a7799bdf3b49d6c91d98", + "last-applied-secret": "dbe414e5bbd906192d94a07211d598cbb4276befe30312deb0732504ffc62d98", "last-applied-secret": strings.Join({ Lifecycle: nil, LivenessProbe: &v1.Probe{ LocalObjectReference: {Name: "auto-some-name-pxc"}, LocalObjectReference: {Name: "some-name-pxc"}, ManagedFields: nil, + ManagedFields: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - Manager: "kube-controller-manager", - Manager: "percona-xtradb-cluster-operator", MinReadySeconds: 0, [mysql] 2026/04/08 14:37:34 packets.go:58 unexpected EOF [mysql] 2026/04/08 14:40:14 packets.go:58 unexpected EOF [mysql] 2026/04/08 14:40:28 packets.go:58 unexpected EOF Name: "auto-config", {Name: "bin", VolumeSource: {EmptyDir: &{}}}, Name: "config", Name: "ist", Name: "mysql", Name: "mysql-admin", Name: "mysql-init-file", Name: "mysql-users-secret-file", Name: "mysqlx", Name: "proxyadm", Namespace: "users-scheduler-10069", Name: "ssl", Name: "ssl-internal", Name: "sst", Name: "stats", {Name: "tmp", VolumeSource: {EmptyDir: &{}}}, Name: "vault-keyring-secret", Name: "write-set", NFS: nil, NodeName: "", NodeSelector: nil, ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "last-applied-secret": "4bcdb6fea9c0299d4331d4c1c4912d8d200ecda89983a7799bdf3b49d6c91d98", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", ...}}, ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "last-applied-secret": "9c37bdb7cce228df3c6dffa335c6ebc05b2269431c8716e28484b4d3290bfdd4", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", ...}}, ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc"}}, ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, ObjectMeta: v1.ObjectMeta{ ObjectMeta: v1.ObjectMeta{ + ObservedGeneration: 0, - ObservedGeneration: 1, - ObservedGeneration: 10, - ObservedGeneration: 2, - ObservedGeneration: 3, - ObservedGeneration: 4, - ObservedGeneration: 5, - ObservedGeneration: 6, - ObservedGeneration: 7, - ObservedGeneration: 8, - ObservedGeneration: 9, - Operation: "Update", - Operation: "Update", Optional: &false, Optional: &true, Optional: &true, Ordinals: nil, OS: nil, Overhead: nil, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "some-name", UID: "d7337649-0424-438f-924f-8b1f0e55dde1", ...}}, OwnerReferences: nil, "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMTVjZWQzZDhlZmFkMmQxMjdjY2U5NWM1ZmQzYWU0NTM2YTYwYjNlNTRjYTJiY2Q2OGRhYTRkYmVhNzc3ZTQwMiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiMTVjZWQzZDhlZmFkMmQxMjdjY2U5NWM1ZmQzYWU0NTM2YTYwYjNlNTRjYTJiY2Q2OGRhYTRkYmVhNzc3ZTQwMiIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNGJjZGI2ZmVhOWMwMjk5ZDQzMzFkNGMxYzQ5MTJkOGQyMDBlY2RhODk5ODNhNzc5OWJkZjNiNDlkNmM5MWQ5OCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzAzMmM2ZWI4ZDBhN2RlMWY5NjgwNWRkMGRiM2ZjYTA4NGFkMDUzNTU4MzViMDZkZjYzMjc3YmI5YmY2NTVjNyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzAzMmM2ZWI4ZDBhN2RlMWY5NjgwNWRkMGRiM2ZjYTA4NGFkMDUzNTU4MzViMDZkZjYzMjc3YmI5YmY2NTVjNyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiOWMzN2JkYjdjY2UyMjhkZjNjNmRmZmEzMzVjNmViYzA1YjIyNjk0MzFjODcxNmUyODQ4NGI0ZDMyOTBiZmRkNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiOWMzN2JkYjdjY2UyMjhkZjNjNmRmZmEzMzVjNmViYzA1YjIyNjk0MzFjODcxNmUyODQ4NGI0ZDMyOTBiZmRkNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiY2Y2NWNmZWNlNWJjMTAzYmRhZTQ2MmI1M2VmZGE4ODE1YjcxNTA2MDJlOTQ3Mjk1NThhNWM1ZDljZTkwOGQwNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiY2Y2NWNmZWNlNWJjMTAzYmRhZTQ2MmI1M2VmZGE4ODE1YjcxNTA2MDJlOTQ3Mjk1NThhNWM1ZDljZTkwOGQwNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZGM2OWE0MWVhMjVmNzBjMmU1ZjY3NjU0YjcyOTg4MWVjNDdkZDk3NDY3ZjU0OWM4NzljNGQ2MTQ5YmZkMGU3NSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNzAzMmM2ZWI4ZDBhN2RlMWY5NjgwNWRkMGRiM2ZjYTA4NGFkMDUzNTU4MzViMDZkZjYzMjc3YmI5YmY2NTVjNyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiY2Y2NWNmZWNlNWJjMTAzYmRhZTQ2MmI1M2VmZGE4ODE1YjcxNTA2MDJlOTQ3Mjk1NThhNWM1ZDljZTkwOGQwNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiY2Y2NWNmZWNlNWJjMTAzYmRhZTQ2MmI1M2VmZGE4ODE1YjcxNTA2MDJlOTQ3Mjk1NThhNWM1ZDljZTkwOGQwNCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZGJlNDE0ZTViYmQ5MDYxOTJkOTRhMDcyMTFkNTk4Y2JiNDI3NmJlZmUzMDMxMmRlYjA3MzI1MDRmZmM2MmQ5OCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZGJlNDE0ZTViYmQ5MDYxOTJkOTRhMDcyMTFkNTk4Y2JiNDI3NmJlZmUzMDMxMmRlYjA3MzI1MDRmZmM2MmQ5OCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZGM2OWE0MWVhMjVmNzBjMmU1ZjY3NjU0YjcyOTg4MWVjNDdkZDk3NDY3ZjU0OWM4NzljNGQ2MTQ5YmZkMGU3NSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiZGM2OWE0MWVhMjVmNzBjMmU1ZjY3NjU0YjcyOTg4MWVjNDdkZDk3NDY3ZjU0OWM4NzljNGQ2MTQ5YmZkMGU3NSIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIy"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJweGMiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJweGMiLCJwZXJjb25hLmNvbS9jb25maWd1cmF0aW9uLWhhc2giOiJkNDFkOGNkOThmMDBiMjA0ZTk4MDA5OThlY2Y4NDI3ZSIsInBlcmNvbmEuY29tL3NzbC1oYXNoIjoiY2ZkNmE1MjM5ODI2ODE3M2I1MWQ3Y2RiMTMzMWMwOWEiLCJwZXJjb25hLmNv"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBw"..., + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJsYXN0LWFwcGxpZWQtc2VjcmV0IjoiNGJjZGI2ZmVhOWMwMjk5ZDQzMzFkNGMxYzQ5MTJkOGQyMDBlY2RhODk5ODNhNzc5OWJkZjNiNDlkNmM5MWQ5OCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6"..., - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InNvbWUtbmFtZSIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9fSwidGVtcGxhdGUiOnsibWV0YWRhdGEiOnsibGFiZWxzIjp7ImFwcC5rdWJlcm5ldGVzLmlvL2NvbXBvbmVudCI6InByb3h5c3FsIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJzb21lLW5hbWUiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifSwiYW5ub3RhdGlvbnMiOnsia3ViZWN0bC5rdWJlcm5ldGVzLmlvL2RlZmF1bHQtY29udGFpbmVyIjoicHJveHlzcWwiLCJwZXJjb25hLmNvbS9jb25maWd1cmF0aW9uLWhhc2giOiJkNDFkOGNkOThmMDBiMjA0ZTk4MDA5OThlY2Y4NDI3ZSIsInBlcmNvbmEuY29tL3NzbC1oYXNoIjoiY2ZkNmE1MjM5ODI2ODE3M2I1MWQ3Y2RiMTMzMWMw"..., "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", "percona.com/ssl-hash": "cfd6a52398268173b51d7cdb1331c09a", "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc", "percona.com/ssl-internal-hash": "751e59a1c1d94ac67f9d6bb6e756acdc", + PeriodSeconds: 0, - PeriodSeconds: 10, + PersistentVolumeClaimRetentionPolicy: nil, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + Phase: "", - Phase: "Pending", + PodManagementPolicy: "", - PodManagementPolicy: "OrderedReady", Ports: []v1.ContainerPort{ PreemptionPolicy: nil, ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}}, + Protocol: "", - Protocol: "TCP", Quobyte: nil, ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...}, + ReadyReplicas: 0, - ReadyReplicas: 1, - ReadyReplicas: 2, - ReadyReplicas: 3, + Replicas: 0, Replicas: &2, - Replicas: 2, - Replicas: &2, + Replicas: &2, Replicas: &3, - Replicas: 3, - Replicas: &3, + Replicas: &3, + ResourceVersion: "", - ResourceVersion: "1775657657967791017", - ResourceVersion: "1775657861376479008", - ResourceVersion: "1775658044277279017", - ResourceVersion: "1775658108650799017", - ResourceVersion: "1775658146912815017", - ResourceVersion: "1775658336834383008", - ResourceVersion: "1775658362371391017", - ResourceVersion: "1775658410938367017", - ResourceVersion: "1775658466265983017", - ResourceVersion: "1775658518103439017", - ResourceVersion: "1775658582421823017", - ResourceVersion: "1775658736456079008", - ResourceVersion: "1775658773786175017", - ResourceVersion: "1775659081975599008", + RestartPolicy: "", - RestartPolicy: "Always", - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, + SchedulerName: "", - SchedulerName: "default-scheduler", SecretName: "internal-some-name", SecretName: "some-name-mysql-init", SecretName: "some-name-ssl", SecretName: "some-name-ssl-internal", SecretName: "some-name-vault", Secret: &v1.SecretVolumeSource{ SecurityContext: nil, Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "some-name", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, SelfLink: "", ServiceAccountName: "default", ServiceName: "some-name-proxysql-unready", ServiceName: "some-name-pxc", SetHostnameAsFQDN: nil, sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 Spec: v1.PersistentVolumeClaimSpec{ Spec: v1.PodSpec{ Spec: v1.StatefulSetSpec{ StartupProbe: nil, Status: v1.PersistentVolumeClaimStatus{ Status: v1.StatefulSetStatus{ StorageClassName: nil, Subdomain: "", - Subresource: "status", SuccessThreshold: 1, Template: v1.PodTemplateSpec{ TerminationGracePeriodSeconds: &30, TerminationGracePeriodSeconds: &600, TerminationGracePeriodSeconds: nil, + TerminationMessagePath: "", - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePolicy: "", - TerminationMessagePolicy: "File", TimeoutSeconds: 5, - Time: s"2026-04-08 14:13:30 +0000 UTC", - Time: s"2026-04-08 14:14:17 +0000 UTC", - Time: s"2026-04-08 14:17:41 +0000 UTC", - Time: s"2026-04-08 14:20:21 +0000 UTC", - Time: s"2026-04-08 14:20:44 +0000 UTC", - Time: s"2026-04-08 14:20:53 +0000 UTC", - Time: s"2026-04-08 14:21:48 +0000 UTC", - Time: s"2026-04-08 14:22:25 +0000 UTC", - Time: s"2026-04-08 14:22:26 +0000 UTC", - Time: s"2026-04-08 14:22:33 +0000 UTC", - Time: s"2026-04-08 14:25:36 +0000 UTC", - Time: s"2026-04-08 14:26:01 +0000 UTC", - Time: s"2026-04-08 14:26:02 +0000 UTC", - Time: s"2026-04-08 14:26:04 +0000 UTC", - Time: s"2026-04-08 14:26:50 +0000 UTC", - Time: s"2026-04-08 14:27:45 +0000 UTC", - Time: s"2026-04-08 14:27:46 +0000 UTC", - Time: s"2026-04-08 14:28:12 +0000 UTC", - Time: s"2026-04-08 14:28:38 +0000 UTC", - Time: s"2026-04-08 14:29:21 +0000 UTC", - Time: s"2026-04-08 14:29:42 +0000 UTC", - Time: s"2026-04-08 14:32:16 +0000 UTC", - Time: s"2026-04-08 14:32:26 +0000 UTC", - Time: s"2026-04-08 14:32:53 +0000 UTC", - Time: s"2026-04-08 14:35:08 +0000 UTC", - Time: s"2026-04-08 14:38:01 +0000 UTC", Tolerations: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, TypeMeta: {}, TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, + UID: "", - UID: "ac9f0b03-a092-4e94-8c26-571d27561a3e", - UID: "e4788fbb-25ac-4f84-bceb-bf1611e846ae", + UpdatedReplicas: 0, - UpdatedReplicas: 1, - UpdatedReplicas: 2, - UpdatedReplicas: 3, + UpdateRevision: "", - UpdateRevision: "some-name-proxysql-55c8f97957", - UpdateRevision: "some-name-proxysql-5857b64f7c", - UpdateRevision: "some-name-proxysql-5d7957976", - UpdateRevision: "some-name-proxysql-6876776859", - UpdateRevision: "some-name-proxysql-849557b8c6", - UpdateRevision: "some-name-proxysql-99559dd8b", - UpdateRevision: "some-name-pxc-7574cf9956", - UpdateRevision: "some-name-pxc-7865f4f8d", - UpdateRevision: "some-name-pxc-7c489f5c8", - UpdateRevision: "some-name-pxc-8565bdfdd", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, &v1.StatefulSet{ VolumeAttributesClassName: nil, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ VolumeDevices: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...}, VolumeName: "", VolumeSource: v1.VolumeSource{ Volumes: []v1.Volume{ VsphereVolume: nil, WorkingDir: "", + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n users-scheduler-10069 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.QDjwiZYIL9 ++ mktemp + local LAST_ERR=/tmp/tmp.zfG11CToZN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QDjwiZYIL9 perconaxtradbcluster.pxc.percona.com "some-name" deleted from users-scheduler-10069 namespace + cat /tmp/tmp.zfG11CToZN + rm /tmp/tmp.QDjwiZYIL9 /tmp/tmp.zfG11CToZN + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.qcYYLdQJfy ++ mktemp + local LAST_ERR=/tmp/tmp.WMRKjk15xv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qcYYLdQJfy No resources found + cat /tmp/tmp.WMRKjk15xv + rm /tmp/tmp.qcYYLdQJfy /tmp/tmp.WMRKjk15xv + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.CYz8kR4Cxy ++ mktemp + local LAST_ERR=/tmp/tmp.juYiNgyDJm + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CYz8kR4Cxy No resources found + cat /tmp/tmp.juYiNgyDJm + rm /tmp/tmp.CYz8kR4Cxy /tmp/tmp.juYiNgyDJm + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.RjCrpdLYLO ++ mktemp + local LAST_ERR=/tmp/tmp.LyR4SHhoLt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RjCrpdLYLO validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.LyR4SHhoLt + rm /tmp/tmp.RjCrpdLYLO /tmp/tmp.LyR4SHhoLt + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace users-scheduler-10069 + rm -rf /tmp/tmp.uv48lQ0yvW + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.D52RF2fw3o + desc 'test passed' ++ mktemp + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.EmvaMcpjua ++ mktemp + local LAST_ERR=/tmp/tmp.I3tVMoH1I5 + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.AiNxY8qsUs + local exit_status=0 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace users-scheduler-10069 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator