Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/logs/pitr-8-0.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 + GTID_PATTERN='[A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + main + create_infra pitr-29966 + local ns=pitr-29966 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch pxc -n pitr-28140 pitr --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/pitr patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.omjFoKbD4U ++ mktemp + local LAST_ERR=/tmp/tmp.8vbQ4ekFLx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.omjFoKbD4U perconaxtradbcluster.pxc.percona.com "pitr" deleted from pitr-28140 namespace + cat /tmp/tmp.8vbQ4ekFLx + rm /tmp/tmp.omjFoKbD4U /tmp/tmp.8vbQ4ekFLx + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.KL2Q6GZpDZ ++ mktemp + local LAST_ERR=/tmp/tmp.wfZ4DpUgSo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KL2Q6GZpDZ perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio" deleted from pitr-28140 namespace + cat /tmp/tmp.wfZ4DpUgSo + rm /tmp/tmp.KL2Q6GZpDZ /tmp/tmp.wfZ4DpUgSo + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.GaqBzs2o1k ++ mktemp + local LAST_ERR=/tmp/tmp.L7ja0IYruX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.GaqBzs2o1k perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio-gtid" deleted from pitr-28140 namespace + cat /tmp/tmp.L7ja0IYruX + rm /tmp/tmp.GaqBzs2o1k /tmp/tmp.L7ja0IYruX + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ awk '-F ' '{print $2}' ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + xargs kubectl delete ns ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.fEQpWHPyK5 ++ mktemp + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + local LAST_ERR=/tmp/tmp.njVb8bKQfa + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.BUSReeO7bq ++ mktemp + local LAST_ERR=/tmp/tmp.LS2kmlRbS8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BUSReeO7bq + cat /tmp/tmp.LS2kmlRbS8 + rm /tmp/tmp.BUSReeO7bq /tmp/tmp.LS2kmlRbS8 + return 0 namespace "cert-manager" deleted namespace "pitr-28140" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fEQpWHPyK5 namespace "pxc-operator" deleted + cat /tmp/tmp.njVb8bKQfa + rm /tmp/tmp.fEQpWHPyK5 /tmp/tmp.njVb8bKQfa + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.pEFqfYdZTJ ++ mktemp + local LAST_ERR=/tmp/tmp.PeXZVgYRnz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pEFqfYdZTJ namespace/pxc-operator created + cat /tmp/tmp.PeXZVgYRnz + rm /tmp/tmp.pEFqfYdZTJ /tmp/tmp.PeXZVgYRnz + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.dEe2cymxit +++ mktemp ++ local LAST_ERR=/tmp/tmp.cKGHzRu7TP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dEe2cymxit ++ cat /tmp/tmp.cKGHzRu7TP ++ rm /tmp/tmp.dEe2cymxit /tmp/tmp.cKGHzRu7TP ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster5 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.8Df2J3oVFC ++ mktemp + local LAST_ERR=/tmp/tmp.wpsFUTpf7T + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster5 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.8Df2J3oVFC Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster5" modified. + cat /tmp/tmp.wpsFUTpf7T + rm /tmp/tmp.8Df2J3oVFC /tmp/tmp.wpsFUTpf7T + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.WaxFfSLSsj ++ mktemp + local LAST_ERR=/tmp/tmp.mJW01HhhQE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WaxFfSLSsj customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.mJW01HhhQE + rm /tmp/tmp.WaxFfSLSsj /tmp/tmp.mJW01HhhQE + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + sed -e 's^namespace: .*^namespace: pxc-operator^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/deploy/cw-rbac.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.Fdu5RCbgKE ++ mktemp + local LAST_ERR=/tmp/tmp.OuF6hXVaVx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Fdu5RCbgKE clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.OuF6hXVaVx + rm /tmp/tmp.Fdu5RCbgKE /tmp/tmp.OuF6hXVaVx + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/deploy/cw-operator.yaml + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + kubectl_bin apply -f - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = ""' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2473-6d392bea^' ++ mktemp + local LAST_OUT=/tmp/tmp.XeHBWjGWEI ++ mktemp + local LAST_ERR=/tmp/tmp.flyni7WP5T + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.XeHBWjGWEI deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.flyni7WP5T + rm /tmp/tmp.XeHBWjGWEI /tmp/tmp.flyni7WP5T + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.bEMqyhvEPx ++ mktemp + local LAST_ERR=/tmp/tmp.iW93b8FWWJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.bEMqyhvEPx pod/percona-xtradb-cluster-operator-55d95dc9d8-j9x6b condition met + cat /tmp/tmp.iW93b8FWWJ E0517 00:36:56.106417 3354 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-j9x6b&resourceVersion=1778978215745350000&timeoutSeconds=493&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" + rm /tmp/tmp.bEMqyhvEPx /tmp/tmp.iW93b8FWWJ + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ head -1 +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wuh4kYJ8Zs +++ mktemp ++ local LAST_ERR=/tmp/tmp.VV3iqQDnHj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Wuh4kYJ8Zs ++ cat /tmp/tmp.VV3iqQDnHj ++ rm /tmp/tmp.Wuh4kYJ8Zs /tmp/tmp.VV3iqQDnHj ++ return 0 + wait_pod percona-xtradb-cluster-operator-55d95dc9d8-j9x6b 480 pxc-operator + local pod=percona-xtradb-cluster-operator-55d95dc9d8-j9x6b + local max_retry=480 + local ns=pxc-operator ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo percona-xtradb-cluster-operator-55d95dc9d8-j9x6b ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-55d95dc9d8-j9x6b condition met E0517 00:37:00.579154 4044 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-j9x6b&resourceVersion=1778978219074881000&timeoutSeconds=472&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-55d95dc9d8-j9x6b to become Ready.Ok + sleep 3 + create_namespace pitr-29966 + local namespace=pitr-29966 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ tail -n1 ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces pitr-29966' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pitr-29966 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pitr-29966 ++ mktemp + local LAST_OUT=/tmp/tmp.08XMLqZSE9 + awk '{print$1}' ++ mktemp + local LAST_ERR=/tmp/tmp.hwKfh7Zd8p + local exit_status=0 ++ seq 0 2 + kubectl_bin get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-29966 ++ mktemp + local LAST_OUT=/tmp/tmp.arZGnaWoSb ++ mktemp + local LAST_ERR=/tmp/tmp.j5gR8Aalxd + local exit_status=0 ++ seq 0 2 + xargs kubectl delete ns + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-29966 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.arZGnaWoSb + cat /tmp/tmp.j5gR8Aalxd + rm /tmp/tmp.arZGnaWoSb /tmp/tmp.j5gR8Aalxd + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-29966 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.08XMLqZSE9 + cat /tmp/tmp.hwKfh7Zd8p Error from server (NotFound): namespaces "pitr-29966" not found + rm /tmp/tmp.08XMLqZSE9 /tmp/tmp.hwKfh7Zd8p + return 1 + : + wait_for_delete namespace/pitr-29966 + local res=namespace/pitr-29966 + echo -n 'waiting for namespace/pitr-29966 to be deleted' waiting for namespace/pitr-29966 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pitr-29966" not found + desc 'create namespace pitr-29966' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pitr-29966 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pitr-29966 ++ mktemp + local LAST_OUT=/tmp/tmp.xy9jp05TPu ++ mktemp + local LAST_ERR=/tmp/tmp.4A42AmGMOi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pitr-29966 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xy9jp05TPu namespace/pitr-29966 created + cat /tmp/tmp.4A42AmGMOi + rm /tmp/tmp.xy9jp05TPu /tmp/tmp.4A42AmGMOi + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.mBi2xXzWkG +++ mktemp ++ local LAST_ERR=/tmp/tmp.cTXHeTPuD2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mBi2xXzWkG ++ cat /tmp/tmp.cTXHeTPuD2 ++ rm /tmp/tmp.mBi2xXzWkG /tmp/tmp.cTXHeTPuD2 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster5 --namespace=pitr-29966 ++ mktemp + local LAST_OUT=/tmp/tmp.ai4D31uHM0 ++ mktemp + local LAST_ERR=/tmp/tmp.uLtBpTlUaJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster5 --namespace=pitr-29966 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ai4D31uHM0 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster5" modified. + cat /tmp/tmp.uLtBpTlUaJ + rm /tmp/tmp.ai4D31uHM0 /tmp/tmp.uLtBpTlUaJ + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.A2xlc18PMM ++ mktemp + local LAST_ERR=/tmp/tmp.hY4NV4UDFD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.A2xlc18PMM secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.hY4NV4UDFD + rm /tmp/tmp.A2xlc18PMM /tmp/tmp.hY4NV4UDFD + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.NxgRE0VqpB ++ mktemp + local LAST_ERR=/tmp/tmp.x3oMdXcziw + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NxgRE0VqpB namespace/cert-manager created + cat /tmp/tmp.x3oMdXcziw + rm /tmp/tmp.NxgRE0VqpB /tmp/tmp.x3oMdXcziw + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.SmkAKAveOE ++ mktemp + local LAST_ERR=/tmp/tmp.AnQw3sZPyT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SmkAKAveOE namespace/cert-manager labeled + cat /tmp/tmp.AnQw3sZPyT + rm /tmp/tmp.SmkAKAveOE /tmp/tmp.AnQw3sZPyT + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.tjhMtkZRea ++ mktemp + local LAST_ERR=/tmp/tmp.wV2Z4uWAwR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tjhMtkZRea namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.wV2Z4uWAwR Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.tjhMtkZRea /tmp/tmp.wV2Z4uWAwR + return 0 + '[' '' == 4.10 ']' + sleep 70 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/issuer.yml ++ mktemp + local LAST_OUT=/tmp/tmp.uUl1Xi4AbD ++ mktemp + local LAST_ERR=/tmp/tmp.avTyGEofpW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/issuer.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uUl1Xi4AbD issuer.cert-manager.io/selfsigning-issuer created + cat /tmp/tmp.avTyGEofpW + rm /tmp/tmp.uUl1Xi4AbD /tmp/tmp.avTyGEofpW + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/cert.yml ++ mktemp + local LAST_OUT=/tmp/tmp.WciTD5knYZ ++ mktemp + local LAST_ERR=/tmp/tmp.L8Z92A4I3x + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/cert.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WciTD5knYZ certificate.cert-manager.io/tls-minio created + cat /tmp/tmp.L8Z92A4I3x Warning: spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. + rm /tmp/tmp.WciTD5knYZ /tmp/tmp.L8Z92A4I3x + return 0 + sleep_with_log 25 + local d=25 + log 'sleeping for 25 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-05-17T00:38:55+0000]' sleeping for 25 seconds [2026-05-17T00:38:55+0000] sleeping for 25 seconds + sleep 25 + start_minio tls-minio + deploy_helm pitr-29966 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=tls-minio + local endpoint=http://minio-service:9000 + minio_args=(--version $MINIO_VER --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set "users[0].accessKey=some-access-key" --set "users[0].secretKey=some-secret-key" --set "users[0].policy=consoleAdmin" --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G) + local minio_args + [[ -n tls-minio ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Sun May 17 00:39:25 2026 NAMESPACE: pitr-29966 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.pitr-29966.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace pitr-29966 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace pitr-29966 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace pitr-29966 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace pitr-29966 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xS24tE9JO8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tdIBuBLK9p ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xS24tE9JO8 ++ cat /tmp/tmp.tdIBuBLK9p ++ rm /tmp/tmp.xS24tE9JO8 /tmp/tmp.tdIBuBLK9p ++ return 0 + MINIO_POD=minio-service-757b74c96c-s6pn8 + wait_pod minio-service-757b74c96c-s6pn8 + local pod=minio-service-757b74c96c-s6pn8 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo minio-service-757b74c96c-s6pn8 ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/minio-service-757b74c96c-s6pn8 condition met E0517 00:40:28.112831 31150 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dminio-service-757b74c96c-s6pn8&resourceVersion=1778978425655616000&timeoutSeconds=347&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/minio-service-757b74c96c-s6pn8 to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.PnxV6KNDR6 ++ mktemp + local LAST_ERR=/tmp/tmp.FB5e3us2gf + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PnxV6KNDR6 /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from pitr-29966 namespace + cat /tmp/tmp.FB5e3us2gf All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: unable to upgrade connection: container aws-cli not found in pod aws-cli_pitr-29966 + rm /tmp/tmp.PnxV6KNDR6 /tmp/tmp.FB5e3us2gf + return 0 + cluster=pitr + mysql_root_pass=root_password + spinup_pxc pitr /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/pitr.yml + local cluster=pitr + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/pitr.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.cmMEJzURuP ++ mktemp + local LAST_ERR=/tmp/tmp.iId9rJxoMr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cmMEJzURuP secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.iId9rJxoMr + rm /tmp/tmp.cmMEJzURuP /tmp/tmp.iId9rJxoMr + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/client.yml + local pvc_name= + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + local LAST_OUT=/tmp/tmp.H5dyLpZiQv + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/client.yml ++ mktemp + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_ERR=/tmp/tmp.w37eItU5r8 + local exit_status=0 + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-29966~ + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2473-6d392bea#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.H5dyLpZiQv deployment.apps/pxc-client created + cat /tmp/tmp.w37eItU5r8 + rm /tmp/tmp.H5dyLpZiQv /tmp/tmp.w37eItU5r8 + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/pitr.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/pitr.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/pitr.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/pitr.yml + local pvc_name= + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' ++ mktemp + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-29966~ + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.PxWesS1bYx + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/pitr.yml + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2473-6d392bea#' ++ mktemp + local LAST_ERR=/tmp/tmp.xcUf5UUlBQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PxWesS1bYx perconaxtradbcluster.pxc.percona.com/pitr created + cat /tmp/tmp.xcUf5UUlBQ + rm /tmp/tmp.PxWesS1bYx /tmp/tmp.xcUf5UUlBQ + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DKkaLCMuyw ++++ mktemp +++ local LAST_ERR=/tmp/tmp.I0P9le99vt +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.DKkaLCMuyw +++ cat /tmp/tmp.I0P9le99vt +++ rm /tmp/tmp.DKkaLCMuyw /tmp/tmp.I0P9le99vt +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0lQa5yZEka ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Ya2HFg2D9L +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.0lQa5yZEka +++ cat /tmp/tmp.Ya2HFg2D9L +++ rm /tmp/tmp.0lQa5yZEka /tmp/tmp.Ya2HFg2D9L +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-29966 ++ mktemp + local LAST_OUT=/tmp/tmp.LjmqqyWwbX ++ mktemp + local LAST_ERR=/tmp/tmp.fDtVwgCYMC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-29966 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-29966 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-29966 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.LjmqqyWwbX + cat /tmp/tmp.fDtVwgCYMC error: no matching resources found + rm /tmp/tmp.LjmqqyWwbX /tmp/tmp.fDtVwgCYMC + return 1 + true + wait_for_running pitr-proxysql 1 + local name=pitr-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ echo pitr-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/pitr-proxysql-0 condition met waiting for pod/pitr-proxysql-0 to become Ready.Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-0 + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-1 ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc pitr -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ base64 --decode ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xNkkYD1W84 +++ mktemp ++ local LAST_ERR=/tmp/tmp.QsVGJ7N1Pq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xNkkYD1W84 ++ cat /tmp/tmp.QsVGJ7N1Pq ++ rm /tmp/tmp.xNkkYD1W84 /tmp/tmp.QsVGJ7N1Pq ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KyWoddq45S +++ mktemp ++ local LAST_ERR=/tmp/tmp.SJnS3txgic ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KyWoddq45S ++ cat /tmp/tmp.SJnS3txgic ++ rm /tmp/tmp.KyWoddq45S /tmp/tmp.SJnS3txgic ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-gd764 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:45:18.721409 6987 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978716004421000&timeoutSeconds=578&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.29x9B2YEnK +++ mktemp ++ local LAST_ERR=/tmp/tmp.uSfJbQvt9u ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.29x9B2YEnK ++ cat /tmp/tmp.uSfJbQvt9u ++ rm /tmp/tmp.29x9B2YEnK /tmp/tmp.uSfJbQvt9u ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-gd764 + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:45:32.382373 8980 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978730189682000&timeoutSeconds=350&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Uy0ZIA2Szh +++ mktemp ++ local LAST_ERR=/tmp/tmp.YZukbsjJ8a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Uy0ZIA2Szh ++ cat /tmp/tmp.YZukbsjJ8a ++ rm /tmp/tmp.Uy0ZIA2Szh /tmp/tmp.YZukbsjJ8a ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-gd764 + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:46:13.459126 14146 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978771578913000&timeoutSeconds=479&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.yb93SkIB9U/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.yb93SkIB9U/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.yb93SkIB9U/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.obTG5jpDm7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PiPEeLx3Dm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.obTG5jpDm7 ++ cat /tmp/tmp.PiPEeLx3Dm ++ rm /tmp/tmp.obTG5jpDm7 /tmp/tmp.PiPEeLx3Dm ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-gd764 + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:46:22.456906 15320 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978780562100000&timeoutSeconds=571&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.yb93SkIB9U/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.yb93SkIB9U/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.yb93SkIB9U/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gXzsp1ADdn +++ mktemp ++ local LAST_ERR=/tmp/tmp.OYcn95oR3B ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gXzsp1ADdn ++ cat /tmp/tmp.OYcn95oR3B ++ rm /tmp/tmp.gXzsp1ADdn /tmp/tmp.OYcn95oR3B ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-gd764 + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:46:30.541093 16419 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978788327300000&timeoutSeconds=560&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.yb93SkIB9U/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.yb93SkIB9U/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.yb93SkIB9U/select-1.sql + is_keyring_plugin_in_use pitr + local cluster=pitr + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + grep -E -o 'early-plugin-load=keyring_\w+.so' + kubectl exec pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + return 1 ++ kubectl_bin get pods --selector=app.kubernetes.io/component=pitr -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LN7KVCMPQK +++ mktemp ++ local LAST_ERR=/tmp/tmp.fDG31OsSMd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/component=pitr -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LN7KVCMPQK ++ cat /tmp/tmp.fDG31OsSMd ++ rm /tmp/tmp.LN7KVCMPQK /tmp/tmp.fDG31OsSMd ++ return 0 + pitr_pod=pitr-pitr-6959bd6dc5-vhqlc + wait_pod pitr-pitr-6959bd6dc5-vhqlc + local pod=pitr-pitr-6959bd6dc5-vhqlc + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-pitr-6959bd6dc5-vhqlc + local container= + set +o xtrace pod/pitr-pitr-6959bd6dc5-vhqlc condition met E0517 00:46:46.508998 18626 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpitr-pitr-6959bd6dc5-vhqlc&resourceVersion=1778978804083425000&timeoutSeconds=550&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pitr-pitr-6959bd6dc5-vhqlc to become ReadyDefaulted container "pitr" out of: pitr, pitr-init (init) .Ok ++ egrep -c 'already exists|is already installed' ++ kubectl_bin logs -l app.kubernetes.io/component=pitr --tail=-1 + [[ 6 == \0 ]] + run_backup pitr on-pitr-minio + local cluster=pitr + local backup=on-pitr-minio + local name= + '[' -z ']' + name=on-pitr-minio + log 'run backup pxc-backup/on-pitr-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-05-17T00:46:54+0000]' run backup pxc-backup/on-pitr-minio [2026-05-17T00:46:54+0000] run backup pxc-backup/on-pitr-minio + yq eval '.metadata.name="on-pitr-minio"' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/on-pitr-minio.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.5g9Z4LXiWo ++ mktemp + local LAST_ERR=/tmp/tmp.ink2451HB5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.5g9Z4LXiWo perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio created + cat /tmp/tmp.ink2451HB5 + rm /tmp/tmp.5g9Z4LXiWo /tmp/tmp.ink2451HB5 + return 0 + wait_backup on-pitr-minio + local backup=on-pitr-minio + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio to reach Succeeded state................Succeeded + write_test_data pitr + local cluster=pitr + local config= + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/conf/client.yml ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.CRmCarFssr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uxkZwaMqbK +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.CRmCarFssr +++ cat /tmp/tmp.uxkZwaMqbK +++ rm /tmp/tmp.CRmCarFssr /tmp/tmp.uxkZwaMqbK +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1DQAsFU3Hc ++++ mktemp +++ local LAST_ERR=/tmp/tmp.sVCvmQURqy +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.1DQAsFU3Hc +++ cat /tmp/tmp.sVCvmQURqy +++ rm /tmp/tmp.1DQAsFU3Hc /tmp/tmp.sVCvmQURqy +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + desc 'write test data' + set +o xtrace ----------------------------------------------------------------------------------- write test data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GgbzG0fa6S +++ mktemp ++ local LAST_ERR=/tmp/tmp.XyDpRTxDgu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GgbzG0fa6S ++ cat /tmp/tmp.XyDpRTxDgu ++ rm /tmp/tmp.GgbzG0fa6S /tmp/tmp.XyDpRTxDgu ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-gd764 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:47:45.565709 26806 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978863646230000&timeoutSeconds=363&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' '-h pitr-proxysql -uroot -proot_password' + local 'command=INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mHuoZ3sp7X +++ mktemp ++ local LAST_ERR=/tmp/tmp.a26t23ZS0E ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mHuoZ3sp7X ++ cat /tmp/tmp.a26t23ZS0E ++ rm /tmp/tmp.mHuoZ3sp7X /tmp/tmp.a26t23ZS0E ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-gd764 + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:47:59.026599 28721 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978876348454000&timeoutSeconds=337&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep_with_log 30 + local d=30 + log 'sleeping for 30 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-05-17T00:48:08+0000]' sleeping for 30 seconds [2026-05-17T00:48:08+0000] sleeping for 30 seconds + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CGAQY8DA6e +++ mktemp ++ local LAST_ERR=/tmp/tmp.ozAAie62ew ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CGAQY8DA6e ++ cat /tmp/tmp.ozAAie62ew ++ rm /tmp/tmp.CGAQY8DA6e /tmp/tmp.ozAAie62ew ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-gd764 + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:48:42.146953 2210 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978920450218000&timeoutSeconds=583&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.yb93SkIB9U/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.yb93SkIB9U/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.yb93SkIB9U/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4pOcRqRmrr +++ mktemp ++ local LAST_ERR=/tmp/tmp.UbSaEONpHs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4pOcRqRmrr ++ cat /tmp/tmp.UbSaEONpHs ++ rm /tmp/tmp.4pOcRqRmrr /tmp/tmp.UbSaEONpHs ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-gd764 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:48:52.635634 3879 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978931588428000&timeoutSeconds=557&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.yb93SkIB9U/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.yb93SkIB9U/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.yb93SkIB9U/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dUVmn16oXl +++ mktemp ++ local LAST_ERR=/tmp/tmp.p93IRPT5nr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dUVmn16oXl ++ cat /tmp/tmp.p93IRPT5nr ++ rm /tmp/tmp.dUVmn16oXl /tmp/tmp.p93IRPT5nr ++ return 0 + client_pod=pxc-client-67fc4995bb-gd764 + wait_pod pxc-client-67fc4995bb-gd764 + local pod=pxc-client-67fc4995bb-gd764 + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-gd764 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:49:06.716913 5686 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778978944602494000&timeoutSeconds=457&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.yb93SkIB9U/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.yb93SkIB9U/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.yb93SkIB9U/select-2.sql ++ is_keyring_plugin_in_use pitr ++ local cluster=pitr ++ [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] ++ grep -E -o 'early-plugin-load=keyring_\w+.so' ++ kubectl exec pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ return 1 + '[' '' ']' + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass1 + patch_secret my-cluster-secrets xtrabackup cGFzczE= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczE= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.e5jS2JcUeL ++ mktemp + local LAST_ERR=/tmp/tmp.2FlZ4osK7M + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.e5jS2JcUeL secret/my-cluster-secrets patched + cat /tmp/tmp.2FlZ4osK7M + rm /tmp/tmp.e5jS2JcUeL /tmp/tmp.2FlZ4osK7M + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K46Qx9xL7E +++ mktemp ++ local LAST_ERR=/tmp/tmp.CwNCP38Zwl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.K46Qx9xL7E ++ cat /tmp/tmp.CwNCP38Zwl ++ rm /tmp/tmp.K46Qx9xL7E /tmp/tmp.CwNCP38Zwl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qWc9UsbpiW +++ mktemp ++ local LAST_ERR=/tmp/tmp.HyXbvW6wgd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qWc9UsbpiW ++ cat /tmp/tmp.HyXbvW6wgd ++ rm /tmp/tmp.qWc9UsbpiW /tmp/tmp.HyXbvW6wgd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HUheVkbEpH +++ mktemp ++ local LAST_ERR=/tmp/tmp.WgLr4uORxy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HUheVkbEpH ++ cat /tmp/tmp.WgLr4uORxy ++ rm /tmp/tmp.HUheVkbEpH /tmp/tmp.WgLr4uORxy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IeTzJaGP39 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZRdCO8aWfs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IeTzJaGP39 ++ cat /tmp/tmp.ZRdCO8aWfs ++ rm /tmp/tmp.IeTzJaGP39 /tmp/tmp.ZRdCO8aWfs ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dWMMnCqYnk +++ mktemp ++ local LAST_ERR=/tmp/tmp.1pJtpzYAga ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dWMMnCqYnk ++ cat /tmp/tmp.1pJtpzYAga ++ rm /tmp/tmp.dWMMnCqYnk /tmp/tmp.1pJtpzYAga ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KJVCtMRmwJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.AD1CIwd8F7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KJVCtMRmwJ ++ cat /tmp/tmp.AD1CIwd8F7 ++ rm /tmp/tmp.KJVCtMRmwJ /tmp/tmp.AD1CIwd8F7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y2KlZzFdIu +++ mktemp ++ local LAST_ERR=/tmp/tmp.bx0xFwdJu4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.y2KlZzFdIu ++ cat /tmp/tmp.bx0xFwdJu4 ++ rm /tmp/tmp.y2KlZzFdIu /tmp/tmp.bx0xFwdJu4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xDjperRNmB +++ mktemp ++ local LAST_ERR=/tmp/tmp.7tzZcb6WkA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xDjperRNmB ++ cat /tmp/tmp.7tzZcb6WkA ++ rm /tmp/tmp.xDjperRNmB /tmp/tmp.7tzZcb6WkA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jZUbLajKjZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.1GAHWTwnX1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jZUbLajKjZ ++ cat /tmp/tmp.1GAHWTwnX1 ++ rm /tmp/tmp.jZUbLajKjZ /tmp/tmp.1GAHWTwnX1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3wHnV25Oim +++ mktemp ++ local LAST_ERR=/tmp/tmp.W96ynd7uoi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3wHnV25Oim ++ cat /tmp/tmp.W96ynd7uoi ++ rm /tmp/tmp.3wHnV25Oim /tmp/tmp.W96ynd7uoi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PubknXAnrR +++ mktemp ++ local LAST_ERR=/tmp/tmp.erJCUPEKrI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PubknXAnrR ++ cat /tmp/tmp.erJCUPEKrI ++ rm /tmp/tmp.PubknXAnrR /tmp/tmp.erJCUPEKrI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3IR0nDhirS +++ mktemp ++ local LAST_ERR=/tmp/tmp.07Wu4Lo6wW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3IR0nDhirS ++ cat /tmp/tmp.07Wu4Lo6wW ++ rm /tmp/tmp.3IR0nDhirS /tmp/tmp.07Wu4Lo6wW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sQTrKyfzv8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.p08uYvJC6o ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sQTrKyfzv8 ++ cat /tmp/tmp.p08uYvJC6o ++ rm /tmp/tmp.sQTrKyfzv8 /tmp/tmp.p08uYvJC6o ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.39NxyJzQIt +++ mktemp ++ local LAST_ERR=/tmp/tmp.ypoUw0yiOm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.39NxyJzQIt ++ cat /tmp/tmp.ypoUw0yiOm ++ rm /tmp/tmp.39NxyJzQIt /tmp/tmp.ypoUw0yiOm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MNphDPhzaB +++ mktemp ++ local LAST_ERR=/tmp/tmp.OWUifypvh1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MNphDPhzaB ++ cat /tmp/tmp.OWUifypvh1 ++ rm /tmp/tmp.MNphDPhzaB /tmp/tmp.OWUifypvh1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Mqd7H2aoYG +++ mktemp ++ local LAST_ERR=/tmp/tmp.b7efN8fEP2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Mqd7H2aoYG ++ cat /tmp/tmp.b7efN8fEP2 ++ rm /tmp/tmp.Mqd7H2aoYG /tmp/tmp.b7efN8fEP2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2O4dmgw0C1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.c9q5FOR5GF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2O4dmgw0C1 ++ cat /tmp/tmp.c9q5FOR5GF ++ rm /tmp/tmp.2O4dmgw0C1 /tmp/tmp.c9q5FOR5GF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SBdCyHl3pg +++ mktemp ++ local LAST_ERR=/tmp/tmp.4iLJoIUml3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SBdCyHl3pg ++ cat /tmp/tmp.4iLJoIUml3 ++ rm /tmp/tmp.SBdCyHl3pg /tmp/tmp.4iLJoIUml3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9Xbu3dDFHv +++ mktemp ++ local LAST_ERR=/tmp/tmp.9Ll1XOh22s ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9Xbu3dDFHv ++ cat /tmp/tmp.9Ll1XOh22s ++ rm /tmp/tmp.9Xbu3dDFHv /tmp/tmp.9Ll1XOh22s ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Qczamct0ep +++ mktemp ++ local LAST_ERR=/tmp/tmp.xlMfneYBYJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Qczamct0ep ++ cat /tmp/tmp.xlMfneYBYJ ++ rm /tmp/tmp.Qczamct0ep /tmp/tmp.xlMfneYBYJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DxPDJoBlFb +++ mktemp ++ local LAST_ERR=/tmp/tmp.cjsV2IaPJW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DxPDJoBlFb ++ cat /tmp/tmp.cjsV2IaPJW ++ rm /tmp/tmp.DxPDJoBlFb /tmp/tmp.cjsV2IaPJW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5txGWOaM1R +++ mktemp ++ local LAST_ERR=/tmp/tmp.gJCsuUxww9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5txGWOaM1R ++ cat /tmp/tmp.gJCsuUxww9 ++ rm /tmp/tmp.5txGWOaM1R /tmp/tmp.gJCsuUxww9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bvzp0GTUyg +++ mktemp ++ local LAST_ERR=/tmp/tmp.UKrL2kGVRU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bvzp0GTUyg ++ cat /tmp/tmp.UKrL2kGVRU ++ rm /tmp/tmp.bvzp0GTUyg /tmp/tmp.UKrL2kGVRU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.g4nkcbbDau +++ mktemp ++ local LAST_ERR=/tmp/tmp.JbhTMfu5sU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.g4nkcbbDau ++ cat /tmp/tmp.JbhTMfu5sU ++ rm /tmp/tmp.g4nkcbbDau /tmp/tmp.JbhTMfu5sU ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZTeWKai8Cb +++ mktemp ++ local LAST_ERR=/tmp/tmp.vQd2b2S4sK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZTeWKai8Cb ++ cat /tmp/tmp.vQd2b2S4sK ++ rm /tmp/tmp.ZTeWKai8Cb /tmp/tmp.vQd2b2S4sK ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.5d0arqph7m ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.bedGboPAj2 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.5d0arqph7m +++++ cat /tmp/tmp.bedGboPAj2 +++++ rm /tmp/tmp.5d0arqph7m /tmp/tmp.bedGboPAj2 +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.ygkWeEbdRa ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.712upCaY7X +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.ygkWeEbdRa +++++ cat /tmp/tmp.712upCaY7X +++++ rm /tmp/tmp.ygkWeEbdRa /tmp/tmp.712upCaY7X +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S0feZfFncG +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lq26HjWQy1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S0feZfFncG ++ cat /tmp/tmp.Lq26HjWQy1 ++ rm /tmp/tmp.S0feZfFncG /tmp/tmp.Lq26HjWQy1 ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass2 + patch_secret my-cluster-secrets xtrabackup cGFzczI= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.F2SU8FRNXj ++ mktemp + local LAST_ERR=/tmp/tmp.K0mXRP87dZ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.F2SU8FRNXj secret/my-cluster-secrets patched + cat /tmp/tmp.K0mXRP87dZ + rm /tmp/tmp.F2SU8FRNXj /tmp/tmp.K0mXRP87dZ + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Mx2ciQJMaI +++ mktemp ++ local LAST_ERR=/tmp/tmp.TaxoVkJ5Px ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Mx2ciQJMaI ++ cat /tmp/tmp.TaxoVkJ5Px ++ rm /tmp/tmp.Mx2ciQJMaI /tmp/tmp.TaxoVkJ5Px ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nFV3Ty5ZQt +++ mktemp ++ local LAST_ERR=/tmp/tmp.pwy1Umc1q5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nFV3Ty5ZQt ++ cat /tmp/tmp.pwy1Umc1q5 ++ rm /tmp/tmp.nFV3Ty5ZQt /tmp/tmp.pwy1Umc1q5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NNpYphR6Cf +++ mktemp ++ local LAST_ERR=/tmp/tmp.kkmfyvLUDU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NNpYphR6Cf ++ cat /tmp/tmp.kkmfyvLUDU ++ rm /tmp/tmp.NNpYphR6Cf /tmp/tmp.kkmfyvLUDU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pp25gG1I8v +++ mktemp ++ local LAST_ERR=/tmp/tmp.yTBNwVlcZd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Pp25gG1I8v ++ cat /tmp/tmp.yTBNwVlcZd ++ rm /tmp/tmp.Pp25gG1I8v /tmp/tmp.yTBNwVlcZd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MFi7PxrocJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.NYhcilbS4v ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MFi7PxrocJ ++ cat /tmp/tmp.NYhcilbS4v ++ rm /tmp/tmp.MFi7PxrocJ /tmp/tmp.NYhcilbS4v ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.urQou4gULL +++ mktemp ++ local LAST_ERR=/tmp/tmp.DbtNW73Smp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.urQou4gULL ++ cat /tmp/tmp.DbtNW73Smp ++ rm /tmp/tmp.urQou4gULL /tmp/tmp.DbtNW73Smp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XAcfWMSllq +++ mktemp ++ local LAST_ERR=/tmp/tmp.8r8cbzHGNI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XAcfWMSllq ++ cat /tmp/tmp.8r8cbzHGNI ++ rm /tmp/tmp.XAcfWMSllq /tmp/tmp.8r8cbzHGNI ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kNQtpiLjZS +++ mktemp ++ local LAST_ERR=/tmp/tmp.DXfBuiz4Tr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kNQtpiLjZS ++ cat /tmp/tmp.DXfBuiz4Tr ++ rm /tmp/tmp.kNQtpiLjZS /tmp/tmp.DXfBuiz4Tr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZOLWyP4drT +++ mktemp ++ local LAST_ERR=/tmp/tmp.mGTFemBrkd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZOLWyP4drT ++ cat /tmp/tmp.mGTFemBrkd ++ rm /tmp/tmp.ZOLWyP4drT /tmp/tmp.mGTFemBrkd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WvcaK6fUdC +++ mktemp ++ local LAST_ERR=/tmp/tmp.1kCLGvtliI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WvcaK6fUdC ++ cat /tmp/tmp.1kCLGvtliI ++ rm /tmp/tmp.WvcaK6fUdC /tmp/tmp.1kCLGvtliI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sBWdJyox6V +++ mktemp ++ local LAST_ERR=/tmp/tmp.AacbGhY1LJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sBWdJyox6V ++ cat /tmp/tmp.AacbGhY1LJ ++ rm /tmp/tmp.sBWdJyox6V /tmp/tmp.AacbGhY1LJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wqr0amdDJB +++ mktemp ++ local LAST_ERR=/tmp/tmp.uRQrIEWKoQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Wqr0amdDJB ++ cat /tmp/tmp.uRQrIEWKoQ ++ rm /tmp/tmp.Wqr0amdDJB /tmp/tmp.uRQrIEWKoQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qWCOoypOrD +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ic7qYGa0gA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qWCOoypOrD ++ cat /tmp/tmp.Ic7qYGa0gA ++ rm /tmp/tmp.qWCOoypOrD /tmp/tmp.Ic7qYGa0gA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uc5U4srSke +++ mktemp ++ local LAST_ERR=/tmp/tmp.Uxo6ZMsVKD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uc5U4srSke ++ cat /tmp/tmp.Uxo6ZMsVKD ++ rm /tmp/tmp.uc5U4srSke /tmp/tmp.Uxo6ZMsVKD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dyDeWbQoPy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Njf79ycMuk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dyDeWbQoPy ++ cat /tmp/tmp.Njf79ycMuk ++ rm /tmp/tmp.dyDeWbQoPy /tmp/tmp.Njf79ycMuk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d5wYIDkUy9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.XZhOACECcM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.d5wYIDkUy9 ++ cat /tmp/tmp.XZhOACECcM ++ rm /tmp/tmp.d5wYIDkUy9 /tmp/tmp.XZhOACECcM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EF7U5cgIFn +++ mktemp ++ local LAST_ERR=/tmp/tmp.cYeNZalquF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EF7U5cgIFn ++ cat /tmp/tmp.cYeNZalquF ++ rm /tmp/tmp.EF7U5cgIFn /tmp/tmp.cYeNZalquF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VTXEs5I44Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.PBcYvjvlYu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VTXEs5I44Q ++ cat /tmp/tmp.PBcYvjvlYu ++ rm /tmp/tmp.VTXEs5I44Q /tmp/tmp.PBcYvjvlYu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N6UPJJ4ztv +++ mktemp ++ local LAST_ERR=/tmp/tmp.CYuK6MUlyF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.N6UPJJ4ztv ++ cat /tmp/tmp.CYuK6MUlyF ++ rm /tmp/tmp.N6UPJJ4ztv /tmp/tmp.CYuK6MUlyF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JG7W3oU4iJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.UiHtWOzXUH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JG7W3oU4iJ ++ cat /tmp/tmp.UiHtWOzXUH ++ rm /tmp/tmp.JG7W3oU4iJ /tmp/tmp.UiHtWOzXUH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bYc7JDQayH +++ mktemp ++ local LAST_ERR=/tmp/tmp.0ii3uVd9IR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bYc7JDQayH ++ cat /tmp/tmp.0ii3uVd9IR ++ rm /tmp/tmp.bYc7JDQayH /tmp/tmp.0ii3uVd9IR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4vPU2mmACl +++ mktemp ++ local LAST_ERR=/tmp/tmp.tS351mdQXB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4vPU2mmACl ++ cat /tmp/tmp.tS351mdQXB ++ rm /tmp/tmp.4vPU2mmACl /tmp/tmp.tS351mdQXB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ynFWgllgGn +++ mktemp ++ local LAST_ERR=/tmp/tmp.o6TeoRU13k ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ynFWgllgGn ++ cat /tmp/tmp.o6TeoRU13k ++ rm /tmp/tmp.ynFWgllgGn /tmp/tmp.o6TeoRU13k ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9BBFhTpELn +++ mktemp ++ local LAST_ERR=/tmp/tmp.z1AV4p3on1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9BBFhTpELn ++ cat /tmp/tmp.z1AV4p3on1 ++ rm /tmp/tmp.9BBFhTpELn /tmp/tmp.z1AV4p3on1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7y6AyPYxvy +++ mktemp ++ local LAST_ERR=/tmp/tmp.UQJuTMml03 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7y6AyPYxvy ++ cat /tmp/tmp.UQJuTMml03 ++ rm /tmp/tmp.7y6AyPYxvy /tmp/tmp.UQJuTMml03 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oXj3mlPJ5b +++ mktemp ++ local LAST_ERR=/tmp/tmp.fXSf3ByfRv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oXj3mlPJ5b ++ cat /tmp/tmp.fXSf3ByfRv ++ rm /tmp/tmp.oXj3mlPJ5b /tmp/tmp.fXSf3ByfRv ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.vhFIRYUCsJ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.0hAg5YckNy +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.vhFIRYUCsJ +++++ cat /tmp/tmp.0hAg5YckNy +++++ rm /tmp/tmp.vhFIRYUCsJ /tmp/tmp.0hAg5YckNy +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Q2Ss8T68EM ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.IEKtT9RAYG +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Q2Ss8T68EM +++++ cat /tmp/tmp.IEKtT9RAYG +++++ rm /tmp/tmp.Q2Ss8T68EM /tmp/tmp.IEKtT9RAYG +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C7d5ix0LqG +++ mktemp ++ local LAST_ERR=/tmp/tmp.kfhS8cquFt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.C7d5ix0LqG ++ cat /tmp/tmp.kfhS8cquFt ++ rm /tmp/tmp.C7d5ix0LqG /tmp/tmp.kfhS8cquFt ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass3 + patch_secret my-cluster-secrets xtrabackup cGFzczM= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczM= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.Bm6fhxwikA ++ mktemp + local LAST_ERR=/tmp/tmp.oVLONrGWSL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Bm6fhxwikA secret/my-cluster-secrets patched + cat /tmp/tmp.oVLONrGWSL + rm /tmp/tmp.Bm6fhxwikA /tmp/tmp.oVLONrGWSL + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GVcuyf28Zj +++ mktemp ++ local LAST_ERR=/tmp/tmp.BKEAlZXt0U ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GVcuyf28Zj ++ cat /tmp/tmp.BKEAlZXt0U ++ rm /tmp/tmp.GVcuyf28Zj /tmp/tmp.BKEAlZXt0U ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2dbofSKNcg +++ mktemp ++ local LAST_ERR=/tmp/tmp.zvpZAG2sz2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2dbofSKNcg ++ cat /tmp/tmp.zvpZAG2sz2 ++ rm /tmp/tmp.2dbofSKNcg /tmp/tmp.zvpZAG2sz2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ekYE3LG7vX +++ mktemp ++ local LAST_ERR=/tmp/tmp.vlSW6ywUMw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ekYE3LG7vX ++ cat /tmp/tmp.vlSW6ywUMw ++ rm /tmp/tmp.ekYE3LG7vX /tmp/tmp.vlSW6ywUMw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VfMEw0YZ10 +++ mktemp ++ local LAST_ERR=/tmp/tmp.xk9r7CwSjw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VfMEw0YZ10 ++ cat /tmp/tmp.xk9r7CwSjw ++ rm /tmp/tmp.VfMEw0YZ10 /tmp/tmp.xk9r7CwSjw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OhgzTLZKE4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qk4MBwjtr9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OhgzTLZKE4 ++ cat /tmp/tmp.Qk4MBwjtr9 ++ rm /tmp/tmp.OhgzTLZKE4 /tmp/tmp.Qk4MBwjtr9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tgfS0qQQBN +++ mktemp ++ local LAST_ERR=/tmp/tmp.xJOiNCK1pn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tgfS0qQQBN ++ cat /tmp/tmp.xJOiNCK1pn ++ rm /tmp/tmp.tgfS0qQQBN /tmp/tmp.xJOiNCK1pn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N3EXEQ1eKS +++ mktemp ++ local LAST_ERR=/tmp/tmp.mLSIQJjzWI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.N3EXEQ1eKS ++ cat /tmp/tmp.mLSIQJjzWI ++ rm /tmp/tmp.N3EXEQ1eKS /tmp/tmp.mLSIQJjzWI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oX1Eb8gbfH +++ mktemp ++ local LAST_ERR=/tmp/tmp.9Waz0LhtSS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oX1Eb8gbfH ++ cat /tmp/tmp.9Waz0LhtSS ++ rm /tmp/tmp.oX1Eb8gbfH /tmp/tmp.9Waz0LhtSS ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5h9ROmhHU9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ag8v9qGbcx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5h9ROmhHU9 ++ cat /tmp/tmp.ag8v9qGbcx ++ rm /tmp/tmp.5h9ROmhHU9 /tmp/tmp.ag8v9qGbcx ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bwtmT6HBz8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.39cbDkNQe5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bwtmT6HBz8 ++ cat /tmp/tmp.39cbDkNQe5 ++ rm /tmp/tmp.bwtmT6HBz8 /tmp/tmp.39cbDkNQe5 ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LeIZD918c4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rEVJA9v8YM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LeIZD918c4 ++ cat /tmp/tmp.rEVJA9v8YM ++ rm /tmp/tmp.LeIZD918c4 /tmp/tmp.rEVJA9v8YM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o6dOlkeoPf +++ mktemp ++ local LAST_ERR=/tmp/tmp.W1PY6YfsBv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o6dOlkeoPf ++ cat /tmp/tmp.W1PY6YfsBv ++ rm /tmp/tmp.o6dOlkeoPf /tmp/tmp.W1PY6YfsBv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MhWqNQE1u2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FlVVi7i7hE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MhWqNQE1u2 ++ cat /tmp/tmp.FlVVi7i7hE ++ rm /tmp/tmp.MhWqNQE1u2 /tmp/tmp.FlVVi7i7hE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.O6nkcHLwf4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.2ROgl81oDw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.O6nkcHLwf4 ++ cat /tmp/tmp.2ROgl81oDw ++ rm /tmp/tmp.O6nkcHLwf4 /tmp/tmp.2ROgl81oDw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bVEZAlIhU4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.soBP00A0aI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bVEZAlIhU4 ++ cat /tmp/tmp.soBP00A0aI ++ rm /tmp/tmp.bVEZAlIhU4 /tmp/tmp.soBP00A0aI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SXm1kUFkyF +++ mktemp ++ local LAST_ERR=/tmp/tmp.fNIe0lye7p ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SXm1kUFkyF ++ cat /tmp/tmp.fNIe0lye7p ++ rm /tmp/tmp.SXm1kUFkyF /tmp/tmp.fNIe0lye7p ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GScOxX5oUm +++ mktemp ++ local LAST_ERR=/tmp/tmp.V5x2wdnnFF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GScOxX5oUm ++ cat /tmp/tmp.V5x2wdnnFF ++ rm /tmp/tmp.GScOxX5oUm /tmp/tmp.V5x2wdnnFF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TXuRG08UT2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.geSTRtrvZn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TXuRG08UT2 ++ cat /tmp/tmp.geSTRtrvZn ++ rm /tmp/tmp.TXuRG08UT2 /tmp/tmp.geSTRtrvZn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IzYJqM3eOx +++ mktemp ++ local LAST_ERR=/tmp/tmp.vWlXEBJm8Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IzYJqM3eOx ++ cat /tmp/tmp.vWlXEBJm8Z ++ rm /tmp/tmp.IzYJqM3eOx /tmp/tmp.vWlXEBJm8Z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UXz7bR7CE3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.RLfvMu76pA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UXz7bR7CE3 ++ cat /tmp/tmp.RLfvMu76pA ++ rm /tmp/tmp.UXz7bR7CE3 /tmp/tmp.RLfvMu76pA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aPepSVnokf +++ mktemp ++ local LAST_ERR=/tmp/tmp.RmDO4mCPBm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aPepSVnokf ++ cat /tmp/tmp.RmDO4mCPBm ++ rm /tmp/tmp.aPepSVnokf /tmp/tmp.RmDO4mCPBm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PkFRW9NvDS +++ mktemp ++ local LAST_ERR=/tmp/tmp.ovHe71DVA2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PkFRW9NvDS ++ cat /tmp/tmp.ovHe71DVA2 ++ rm /tmp/tmp.PkFRW9NvDS /tmp/tmp.ovHe71DVA2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zk0cEEr8oB +++ mktemp ++ local LAST_ERR=/tmp/tmp.nM2kJjUKWM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Zk0cEEr8oB ++ cat /tmp/tmp.nM2kJjUKWM ++ rm /tmp/tmp.Zk0cEEr8oB /tmp/tmp.nM2kJjUKWM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XwY3pzO36C +++ mktemp ++ local LAST_ERR=/tmp/tmp.Yx06Vw1uz1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XwY3pzO36C ++ cat /tmp/tmp.Yx06Vw1uz1 ++ rm /tmp/tmp.XwY3pzO36C /tmp/tmp.Yx06Vw1uz1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y2ccVsJZSm +++ mktemp ++ local LAST_ERR=/tmp/tmp.JIWfaJs6xr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y2ccVsJZSm ++ cat /tmp/tmp.JIWfaJs6xr ++ rm /tmp/tmp.Y2ccVsJZSm /tmp/tmp.JIWfaJs6xr ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7qdGfvotip +++ mktemp ++ local LAST_ERR=/tmp/tmp.RoSnFsHh2p ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7qdGfvotip ++ cat /tmp/tmp.RoSnFsHh2p ++ rm /tmp/tmp.7qdGfvotip /tmp/tmp.RoSnFsHh2p ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.TDH08eJDOn ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.P4PPEjRsxb +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.TDH08eJDOn +++++ cat /tmp/tmp.P4PPEjRsxb +++++ rm /tmp/tmp.TDH08eJDOn /tmp/tmp.P4PPEjRsxb +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Qd0DA8NdN6 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.dAqwvoOnoR +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Qd0DA8NdN6 +++++ cat /tmp/tmp.dAqwvoOnoR +++++ rm /tmp/tmp.Qd0DA8NdN6 /tmp/tmp.dAqwvoOnoR +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.55DdxHCVSF +++ mktemp ++ local LAST_ERR=/tmp/tmp.OPcwG5nPe3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.55DdxHCVSF ++ cat /tmp/tmp.OPcwG5nPe3 ++ rm /tmp/tmp.55DdxHCVSF /tmp/tmp.OPcwG5nPe3 ++ return 0 + [[ 2 == \2 ]] + echo + compare_kubectl secret/pitr-mysql-init + local resource=secret/pitr-mysql-init + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml + local new_result=/tmp/tmp.yb93SkIB9U/secret_pitr-mysql-init.yml + desc 'compare secret/pitr-mysql-init-' + set +o xtrace ----------------------------------------------------------------------------------- compare secret/pitr-mysql-init- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.33 >= 1.33' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-k133.yml ']' + version_gt 1.29 ++ echo '1.33 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129.yml ']' + version_gt 1.27 ++ bc -l ++ echo '1.33 >= 1.27' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.33 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-k124.yml ']' + version_gt 1.22 ++ echo '1.33 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.33 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-oc.yml ']' + version_gt 1.29 ++ echo '1.33 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | del(.spec.updateStrategy.rollingUpdate.maxUnavailable) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("pitr-29966", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml secret/pitr-mysql-init ++ mktemp + local LAST_OUT=/tmp/tmp.wdrtFEHMtU ++ mktemp + local LAST_ERR=/tmp/tmp.24meh13dOP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml secret/pitr-mysql-init + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wdrtFEHMtU + cat /tmp/tmp.24meh13dOP + rm /tmp/tmp.wdrtFEHMtU /tmp/tmp.24meh13dOP + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml /tmp/tmp.yb93SkIB9U/secret_pitr-mysql-init.yml + log 'compare_kubectl: secret/pitr-mysql-init OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-05-17T00:58:10+0000]' compare_kubectl: secret/pitr-mysql-init OK [2026-05-17T00:58:10+0000] compare_kubectl: secret/pitr-mysql-init OK + mysql_host=pitr-pxc-0.pitr-pxc.pitr-29966 ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-29966 ++ local host=pitr-pxc-0.pitr-pxc.pitr-29966 +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-29966 -uroot -proot_password' +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-29966 -uroot -proot_password' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.GPc8FoWBx4 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.NYLhmkWatu ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.GPc8FoWBx4 ++++ cat /tmp/tmp.NYLhmkWatu ++++ rm /tmp/tmp.GPc8FoWBx4 /tmp/tmp.NYLhmkWatu ++++ return 0 +++ client_pod=pxc-client-67fc4995bb-gd764 +++ wait_pod pxc-client-67fc4995bb-gd764 +++ local pod=pxc-client-67fc4995bb-gd764 +++ local max_retry=480 +++ local ns= ++++ grep -E '^(pxc|proxysql)$' ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ echo pxc-client-67fc4995bb-gd764 +++ local container= +++ set +o xtrace pod/pxc-client-67fc4995bb-gd764 condition met E0517 00:58:13.927417 13606 reflector.go:227] "Failed to watch" err="Get \"https://35.254.121.66/api/v1/namespaces/pitr-29966/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-gd764&resourceVersion=1778979491784482000&timeoutSeconds=511&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-gd764 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local gtid=1bbfc759-5189-11f1-93b8-dbf58450df6f:29 ++ [[ ! 1bbfc759-5189-11f1-93b8-dbf58450df6f:29 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo 1bbfc759-5189-11f1-93b8-dbf58450df6f:29 + gtid=1bbfc759-5189-11f1-93b8-dbf58450df6f:29 + log '[CASE 1]: gtid_executed: 1bbfc759-5189-11f1-93b8-dbf58450df6f:29' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-05-17T00:58:21+0000]' '[CASE' '1]:' gtid_executed: 1bbfc759-5189-11f1-93b8-dbf58450df6f:29 [2026-05-17T00:58:21+0000] [CASE 1]: gtid_executed: 1bbfc759-5189-11f1-93b8-dbf58450df6f:29 + write_data_for_pitr pitr 100503 3 + local cluster=pitr + local start=100503 + local rows=3 ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pCBfoZRDHS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yqZkLMKfKf +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.pCBfoZRDHS +++ cat /tmp/tmp.yqZkLMKfKf +++ rm /tmp/tmp.pCBfoZRDHS /tmp/tmp.yqZkLMKfKf +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.jXcbxPL9Fg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ex7hPV4707 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.jXcbxPL9Fg +++ cat /tmp/tmp.ex7hPV4707 +++ rm /tmp/tmp.jXcbxPL9Fg /tmp/tmp.ex7hPV4707 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + local sql= + (( i=0 )) + (( i//g' + /usr/bin/sed -e 's//1bbfc759-5189-11f1-93b8-dbf58450df6f:29/g' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-29966~ + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/pitr/conf/restore-on-pitr-minio-gtid.yaml + kubectl_bin apply -f - + /usr/bin/sed -e 's///g' ++ mktemp + local LAST_OUT=/tmp/tmp.yXTyyC1nAR ++ mktemp + local LAST_ERR=/tmp/tmp.SeonoBhvWJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yXTyyC1nAR perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-gtid created + cat /tmp/tmp.SeonoBhvWJ + rm /tmp/tmp.yXTyyC1nAR /tmp/tmp.SeonoBhvWJ + return 0 + wait_backup_restore restore-on-pitr-minio-gtid 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Stopping Cluster state 2026-05-17T00:59:51 pxc-restore/restore-on-pitr-minio-gtid state: Stopping Cluster + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace .Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace ..Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace .....Error from server (NotFound): pods "pitr-pxc-1" not found + wait_for_delete pod/pitr-pxc-0 + local res=pod/pitr-pxc-0 + echo -n 'waiting for pod/pitr-pxc-0 to be deleted' waiting for pod/pitr-pxc-0 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-0" not found + wait_backup_restore restore-on-pitr-minio-gtid Restoring + local backup_name=restore-on-pitr-minio-gtid + local target_state=Restoring + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Restoring state 2026-05-17T01:00:35 pxc-restore/restore-on-pitr-minio-gtid state: Restoring + wait_backup_restore restore-on-pitr-minio-gtid 'Preparing Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Preparing Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Preparing Cluster state 2026-05-17T01:00:38 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:00:41 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:00:44 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:00:47 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:00:50 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:00:53 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:00:55 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:00:59 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:01:01 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:01:04 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-05-17T01:01:07 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster + wait_backup_restore restore-on-pitr-minio-gtid 'Point-in-time recovering' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Point-in-time recovering' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Point-in-time recovering state 2026-05-17T01:01:10 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:13 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:16 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:20 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:24 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:28 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:32 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:36 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:39 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:43 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:46 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:49 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:52 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:55 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:01:58 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:02 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:05 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:07 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:10 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:13 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:17 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:20 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:23 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:25 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:27 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:31 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:35 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:40 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:44 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:48 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:53 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:02:57 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:03:01 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:03:05 pxc-restore/restore-on-pitr-minio-gtid state: Preparing Cluster 2026-05-17T01:03:09 pxc-restore/restore-on-pitr-minio-gtid state: Point-in-time recovering + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ........................................................................................................................2026-05-17T00:36:44.804Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.33.11-gke.1137000"} 2026-05-17T00:36:44.804Z INFO setup Manager starting up {"gitCommit": "6d392bea35083cf1f9dca50d602c31a2f57502ed", "gitBranch": "PR-2473-6d392bea", "buildTime": "2026-05-16T22:17:40Z", "goVersion": "go1.26.3", "os": "linux", "arch": "amd64"} 2026-05-17T00:36:44.804Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "", "enabled": ""} 2026-05-17T00:36:44.808Z INFO setup Registering Components. 2026-05-17T00:36:45.608Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-05-17T00:36:45.608Z INFO setup Starting the Cmd. 2026-05-17T00:36:45.608Z INFO controller-runtime.metrics Starting metrics server 2026-05-17T00:36:45.608Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-05-17T00:36:45.608Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-05-17T00:36:45.608Z INFO controller-runtime.webhook Starting webhook server 2026-05-17T00:36:45.608Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-05-17T00:36:45.608Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-05-17T00:36:45.608Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-05-17T00:36:45.709Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-05-17T00:36:45.739Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-05-17T00:36:45.739Z DEBUG events percona-xtradb-cluster-operator-55d95dc9d8-j9x6b_e2898efd-6cb9-483b-b2b4-82e9c7dc3030 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"0a8d17ec-9951-465f-a869-ad329ab0187f","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1778978205731471009"}, "reason": "LeaderElection"} 2026-05-17T00:36:45.739Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-05-17T00:36:45.739Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-05-17T00:36:45.739Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-05-17T00:36:45.740Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-05-17T00:36:45.940Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-05-17T00:36:45.940Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-05-17T00:36:45.940Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-05-17T00:36:45.940Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-05-17T00:36:45.940Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-05-17T00:36:45.940Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-05-17T00:40:48.793Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "version": "1.20.0"} 2026-05-17T00:40:55.318Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "auto-pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-05-17T00:40:55.468Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-05-17T00:40:55.632Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-05-17T00:40:55.788Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-05-17T00:40:55.948Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-05-17T00:40:55.981Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "pitr-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-05-17T00:40:56.035Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-05-17T00:40:56.201Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c6057b2a-327a-4b42-bdf8-78bf045fe834", "object": "pitr-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-05-17T00:40:56.496Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "86f94532-08ed-4345-a044-1c0f09a39337", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-05-17T00:40:56.516Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "86f94532-08ed-4345-a044-1c0f09a39337", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-05-17T00:42:12.542Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6", "user": "operator"} 2026-05-17T00:42:12.573Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6", "user": "monitor"} 2026-05-17T00:42:12.619Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6"} 2026-05-17T00:42:12.654Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6"} 2026-05-17T00:42:12.690Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6", "user": "xtrabackup"} 2026-05-17T00:42:12.728Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6"} 2026-05-17T00:42:12.759Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6", "user": "replication"} 2026-05-17T00:42:12.768Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "efddd218-9c99-441b-84cb-8a2631cd3da6", "err": "get primary pxc pod: not found"} 2026-05-17T00:42:17.862Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1fd5456c-7764-4c5b-a0cb-ffc9e13beb18", "err": "get primary pxc pod: not found"} 2026-05-17T00:42:22.987Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f1c4f10a-7d13-4455-9705-ff34826c87f3", "err": "get primary pxc pod: not found"} 2026-05-17T00:42:28.084Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "3b4e6937-f6c0-4be4-a790-769e9076f5ea", "err": "get primary pxc pod: not found"} 2026-05-17T00:44:53.953Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "75fc0d13-05e4-47d1-a6c4-20c8497eee24", "user": "root"} 2026-05-17T00:44:54.057Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "75fc0d13-05e4-47d1-a6c4-20c8497eee24", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}"} 2026-05-17T00:44:54.130Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "75fc0d13-05e4-47d1-a6c4-20c8497eee24", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-05-17T00:44:54.186Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "75fc0d13-05e4-47d1-a6c4-20c8497eee24", "new version": "8.0.43-34.1"} 2026-05-17T00:44:54.849Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "42bd50cb-c6b0-430c-a90d-d5184499f26b", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}", "hashChanged": false, "metaChanged": true}   &v1.Service{    TypeMeta: v1.TypeMeta{ -  Kind: "Service", +  Kind: "", -  APIVersion: "v1", +  APIVersion: "",    },    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "3544a1e9-b585-497d-ac01-ec912603d698", +  UID: "", -  ResourceVersion: "1778978694122159007", +  ResourceVersion: "",    Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:44:54 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil,    Annotations: map[string]string{ -  "cloud.google.com/neg": `{"ingress":true}`,    "percona.com/last-config-hash": "eyJwb3J0cyI6W3sibmFtZSI6Imh0dHAiLCJwb3J0Ijo4MDgwLCJ0YXJnZXRQb3J0"...,    },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "v1", -  Time: s"2026-05-17 00:44:54 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.ServiceSpec{    Ports: []v1.ServicePort{    {    Name: "http", -  Protocol: "TCP", +  Protocol: "",    AppProtocol: nil,    Port: 8080,    TargetPort: intstr.IntOrString{    Type: 0, -  IntVal: 8080, +  IntVal: 0,    StrVal: "",    },    NodePort: 0,    },    },    Selector: {"app.kubernetes.io/component": "pitr", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    ClusterIP: "34.118.232.171", -  ClusterIPs: []string{"34.118.232.171"}, +  ClusterIPs: nil,    Type: "ClusterIP",    ExternalIPs: nil, -  SessionAffinity: "None", +  SessionAffinity: "",    LoadBalancerIP: "",    LoadBalancerSourceRanges: nil,    ... // 3 identical fields    PublishNotReadyAddresses: false,    SessionAffinityConfig: nil, -  IPFamilies: []v1.IPFamily{"IPv4"}, +  IPFamilies: nil, -  IPFamilyPolicy: &"SingleStack", +  IPFamilyPolicy: nil,    AllocateLoadBalancerNodePorts: nil,    LoadBalancerClass: nil, -  InternalTrafficPolicy: &"Cluster", +  InternalTrafficPolicy: nil,    TrafficDistribution: nil,    },    Status: {},   } 2026-05-17T00:44:56.347Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "75fc0d13-05e4-47d1-a6c4-20c8497eee24"} 2026-05-17T00:45:02.646Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "a21fc4af-fe0f-4b27-ac04-2170dc0500dd"} 2026-05-17T00:45:07.467Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c32c13d2-dc75-42dd-9e0c-9e79a3a6b6bf"} 2026-05-17T00:45:12.958Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "e7f1fe71-a879-45b4-be02-e158c7286ee2"} 2026-05-17T00:45:18.120Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "a20cebea-0e71-4a4d-86b4-c40b2ce73ca8"} 2026-05-17T00:45:23.265Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "8ce4945e-72b8-4090-8a11-dd2fcb533167"} 2026-05-17T00:45:28.558Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "15c5f1a5-e141-4c3b-87aa-df9f8462f0a5"} 2026-05-17T00:45:34.036Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "99a641cc-d75f-4bee-8f17-71ac89975591"} 2026-05-17T00:45:39.336Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "3122d9bd-e793-4ca0-a4b2-ae4cde29bcb8"} 2026-05-17T00:45:44.642Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "08643aa5-204c-4baf-a6d6-a96d03cf8564"} 2026-05-17T00:45:49.859Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "524810cd-7ce9-4b2e-b0c4-9c332b601038"} 2026-05-17T00:45:55.236Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "508c1feb-3f0d-4c56-90ec-30e03e08b073"} 2026-05-17T00:46:00.141Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c8df3f76-39ed-4657-a01b-a3da60985413"} 2026-05-17T00:46:05.629Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "fe1006a4-0006-4a7e-becd-55f0aaa595b4"} 2026-05-17T00:46:11.344Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "fa3ea564-8960-429a-9dc0-c828eb8d8a50"} 2026-05-17T00:46:16.142Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0b304087-4988-40f7-8d06-7a3566620cd1"} 2026-05-17T00:46:20.934Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "5c4b375b-14ac-45b0-ac2e-5c49c8a502e8"} 2026-05-17T00:46:26.558Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f5a7719b-9a74-4f2d-bdd8-584e81adbfcc"} 2026-05-17T00:46:31.846Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "6f966eab-7543-4c15-913e-536fd0a895dc"} 2026-05-17T00:46:37.059Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "60dba711-33e7-420b-8fea-aeeccfd19828"} 2026-05-17T00:46:42.045Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f1521523-0f4f-4f6d-b79b-b21701dc2142"} 2026-05-17T00:46:47.243Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "325c1a39-1bb7-4707-88dd-fce7038a9bdd"} 2026-05-17T00:46:52.942Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "66bfb647-ac5a-4dee-bf92-dea6c5642205"} 2026-05-17T00:46:57.935Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "89e4e326-f395-490a-b24a-69761f18f38e"} 2026-05-17T00:46:59.088Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "d5adeedc-895e-45f9-8397-5fd456d2195a", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:46:59.122Z INFO Created a new backup job {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "d5adeedc-895e-45f9-8397-5fd456d2195a", "namespace": "pitr-29966", "name": "xb-on-pitr-minio"} 2026-05-17T00:47:03.245Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "8d532b19-74a9-4b50-90d1-b7f8ee5a4086"} 2026-05-17T00:47:04.123Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "9b0fc750-1e26-4c01-a308-3dec7bd0422c", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:04.183Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "eaa0db19-eec5-48ec-867b-8bbb10e27d22", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:08.662Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "442c4d32-1427-4326-82ca-20f8970973e9"} 2026-05-17T00:47:09.217Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "233b2587-48e9-41ea-a11d-a68d737f8a51", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:13.736Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "618c5f91-f548-4885-8356-13e99de8cca3"} 2026-05-17T00:47:14.262Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "b2357197-f678-48c5-a4b1-c947cbc06a87", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:18.952Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0683a193-5bc0-4495-810c-c573b96b6e5a"} 2026-05-17T00:47:19.304Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "1bcc087d-042e-4b67-aed0-0e1f2469429d", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:24.154Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0c588003-6ad7-4018-aa66-97955d31995f"} 2026-05-17T00:47:24.344Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "ebda350d-77b5-4c27-a315-ada7f7b902d5", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:29.386Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "ffc3262e-2de2-4668-b6c6-475156cdd674", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:29.451Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "91d42192-9434-48eb-8e79-4f45910a8c87", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:29.656Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "8c563368-fec0-4360-9ff7-5aff967bb219"} 2026-05-17T00:47:34.489Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "cda5983b-dfd0-47b9-9a44-edc3207ef4a8", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-05-17T00:47:34.529Z INFO Backup succeeded {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "cda5983b-dfd0-47b9-9a44-edc3207ef4a8", "job": "xb-on-pitr-minio"} 2026-05-17T00:47:34.529Z DEBUG Removing binlog gap file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "cda5983b-dfd0-47b9-9a44-edc3207ef4a8", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-6959bd6dc5-vhqlc"} 2026-05-17T00:47:34.948Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "16d2dfd6-008e-45ec-9a53-b0746e90ca23"} 2026-05-17T00:47:35.746Z DEBUG Removing binlog timeline file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "cda5983b-dfd0-47b9-9a44-edc3207ef4a8", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-6959bd6dc5-vhqlc"} 2026-05-17T00:47:35.816Z DEBUG Removing mysql-init secret {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "on-pitr-minio", "reconcileID": "cda5983b-dfd0-47b9-9a44-edc3207ef4a8", "job": "xb-on-pitr-minio", "secret": "pitr-mysql-init"} 2026-05-17T00:47:40.248Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "e640dfa8-11cb-4a7c-9aa7-bbc213415ac4"} 2026-05-17T00:47:45.965Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "4fb4597a-ba8a-4617-934b-1124e06310d6"} 2026-05-17T00:47:51.830Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "3145c5c7-bb2f-47d1-a066-954f9e767f56"} 2026-05-17T00:47:57.537Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "93b668d4-da5e-4d7b-9c5a-4bc0b7598663"} 2026-05-17T00:48:03.355Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "e411166a-04d5-4678-ac39-c65e836b9dad"} 2026-05-17T00:48:08.756Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "5a5af1fe-d898-4c86-a268-5cfce9543789"} 2026-05-17T00:48:14.357Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "e9179df0-407b-4be5-b4fd-feeb805d5971"} 2026-05-17T00:48:20.166Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "106588b8-fe8e-4d3d-bce1-e9babfa479ae"} 2026-05-17T00:48:26.258Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "fbc545d1-f848-44b7-8785-198ed75dd322"} 2026-05-17T00:48:32.320Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "6aa17d7f-bba4-4fe1-b90f-7d0327304bb6"} 2026-05-17T00:48:38.128Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "3803fe4b-bfa4-4c91-8681-be055127e7cf"} 2026-05-17T00:48:43.943Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "33418903-9579-4f3b-bf7b-3c70d093451b"} 2026-05-17T00:48:48.574Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d3ce28cc-9019-425a-b88d-a497336be620", "latest": "2026-05-17 00:48:46 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-05-17T00:48:49.275Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d3ce28cc-9019-425a-b88d-a497336be620"} 2026-05-17T00:48:55.928Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "db77f7dc-4a5f-4950-b268-aa93be86cfe5"} 2026-05-17T00:49:01.221Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "5c824430-80a1-4424-b943-91c6faca34bd"} 2026-05-17T00:49:06.555Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "4ba0325f-7edd-4bcd-a34f-8335ecc697f5"} 2026-05-17T00:49:12.269Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "87c6c377-499f-49c7-af35-02d246edfea6"} 2026-05-17T00:49:17.124Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f6eb94e6-7831-4c53-988e-dd01f9f9e2c1", "user": "xtrabackup"} 2026-05-17T00:49:17.139Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f6eb94e6-7831-4c53-988e-dd01f9f9e2c1", "user": "xtrabackup"} 2026-05-17T00:49:17.156Z INFO MySQL init secret created {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f6eb94e6-7831-4c53-988e-dd01f9f9e2c1", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-05-17T00:49:17.176Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f6eb94e6-7831-4c53-988e-dd01f9f9e2c1", "user": "xtrabackup"} 2026-05-17T00:49:17.189Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f6eb94e6-7831-4c53-988e-dd01f9f9e2c1", "user": "xtrabackup"} 2026-05-17T00:49:17.195Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f6eb94e6-7831-4c53-988e-dd01f9f9e2c1", "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258"} 2026-05-17T00:49:17.198Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f6eb94e6-7831-4c53-988e-dd01f9f9e2c1", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "0e68f460-9e35-45e0-a3fb-31d884297194", +  UID: "", -  ResourceVersion: "1778978691753151022", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiI5MGQwNDg1Y2ZkZjc2OWIxZjZkOGVlMDYxY2JlNTIzZSIsInBlcmNvbmEuY29tL3NzbC1pbnRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:40:55 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:44:51 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc", +  "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258",    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e",    "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-5747c6c878", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-5747c6c878", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-05-17T00:49:18.954Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "3c4c583b-514a-4eed-94a0-6f3a62d16629"} 2026-05-17T00:50:10.671Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "cb3ac31f-b4a4-4833-884e-4bba92fb1771", "latest": "2026-05-17 00:50:07 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-05-17T00:50:22.273Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f63c733a-c578-47b1-9025-4d49066411b4", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T00:51:14.252Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "b92fa051-6308-4773-9a95-b133fbd6ebc8", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T00:51:19.956Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "3e9e6f30-6d66-4993-8516-16eccc7e46be", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:51:25.603Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "7b4f7ad8-0605-4ef9-b85b-babc587b8b2d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:51:31.297Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "73f0fb81-d9d2-4d10-8104-6368e4999946", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:51:36.942Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "228bd0a9-10da-4ffb-92af-ea648a3810ef", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:51:42.604Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "afbb3f95-b79b-4129-bd4c-668dbe793503", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:51:48.071Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "4efe460d-a690-4dca-a6f2-eb2f0f578f74", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:51:53.526Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "27b73c0d-d392-4ba5-946f-922a5256c0e4", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:52:01.076Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0db39c88-15c5-4d30-9302-ba1144b441f5", "latest": "2026-05-17 00:50:23 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-05-17T00:52:02.841Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0db39c88-15c5-4d30-9302-ba1144b441f5"} 2026-05-17T00:52:07.841Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1fae5529-f887-4a15-a067-4b1ae6f0bfdc"} 2026-05-17T00:52:12.785Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1b07aeac-19f8-460c-b54d-f5ae4d15e91b", "user": "xtrabackup"} 2026-05-17T00:52:12.803Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1b07aeac-19f8-460c-b54d-f5ae4d15e91b", "user": "xtrabackup"} 2026-05-17T00:52:12.827Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1b07aeac-19f8-460c-b54d-f5ae4d15e91b", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-05-17T00:52:12.864Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1b07aeac-19f8-460c-b54d-f5ae4d15e91b", "user": "xtrabackup"} 2026-05-17T00:52:12.878Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1b07aeac-19f8-460c-b54d-f5ae4d15e91b", "user": "xtrabackup"} 2026-05-17T00:52:12.883Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1b07aeac-19f8-460c-b54d-f5ae4d15e91b", "last-applied-secret": "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d"} 2026-05-17T00:52:12.889Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1b07aeac-19f8-460c-b54d-f5ae4d15e91b", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "0e68f460-9e35-45e0-a3fb-31d884297194", +  UID: "", -  ResourceVersion: "1778979115516319022", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:49:17 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:51:55 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258", +  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e",    "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-7ff859f9cc", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-7ff859f9cc", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-05-17T00:52:14.042Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0aff2c7c-34ce-4d30-b48d-3ae9058470e5"} 2026-05-17T00:53:02.401Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "61a1d473-8354-4237-b8fa-b601e62841e7", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:53:04.601Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "1ae78bc9-720d-4f95-b77f-2cb726eb86af", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:53:07.033Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "589c9434-8c4d-4998-9517-52995693bbb0", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:53:15.377Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "6b87fb8f-96cb-46b7-bb63-3c75f0e9e3fc", "latest": "2026-05-17 00:53:11 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-05-17T00:53:27.174Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "2e9ec981-8650-4b59-9e01-38c1c5f5cce2", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T00:54:19.430Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "9e7ed862-a66c-4e2b-a067-898280deb129", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T00:54:25.130Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "e816ad67-d0e6-4acf-b604-632e5762b440", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:54:30.875Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "2326eb90-0029-49d8-968d-65bc999806f7", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:54:36.536Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "fd9dc5f1-a4c5-4053-b1ea-d2ef19d5f126", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:54:42.182Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c92b3449-59d1-4376-9ad9-4a2dab8bf73d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:54:47.824Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "903dc2c2-081f-4870-ac11-88abc5faa891", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:54:53.288Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f0e280e2-9eea-4f99-9a6d-6fbe159eb4e8", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:54:53.990Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f65ade4f-f9a4-4bdb-8f4e-dc1cee19aff3", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:54:59.642Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "885948d8-cb94-445d-8797-c38b64ad219b", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:55:08.839Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "547ac997-1277-4db7-82e9-44ec8cb0f9c0"} 2026-05-17T00:55:14.048Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "12b31ba8-4551-49fb-8fbb-0d89722a3e3d"} 2026-05-17T00:55:16.920Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10", "user": "xtrabackup"} 2026-05-17T00:55:16.937Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10", "user": "xtrabackup"} 2026-05-17T00:55:16.956Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-05-17T00:55:16.974Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10", "user": "xtrabackup"} 2026-05-17T00:55:16.991Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10", "user": "xtrabackup"} 2026-05-17T00:55:16.997Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27"} 2026-05-17T00:55:17.000Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "0e68f460-9e35-45e0-a3fb-31d884297194", +  UID: "", -  ResourceVersion: "1778979303531567022", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiJkZjhjYmI5YmQ2NzM1MmMxMDIyOGY0ODcwNjkwZmU4MDIwYzE5MzNjNjkxMWY3Yjk0NmM0N2UyYjc2MDgwYjI3IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:52:12 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:55:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d", +  "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e",    "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-5bc96ddc67", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-5bc96ddc67", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-05-17T00:55:19.361Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "514ca6d2-2c70-4d66-8e68-8fdc1ea99b10"} 2026-05-17T00:56:05.988Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "49742836-f4d0-4e61-bf53-055e390ef059", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:08.363Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "92144c0b-3608-48e3-8391-0a1d11567829", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:11.107Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "b985d5c3-813b-41fa-95dd-139b0065e95d", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:13.527Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d1dac5cd-059b-4442-9dcc-a4803c1d5992", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:16.176Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0ae38a50-268d-4eb2-879b-93e22bf265f4", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:18.526Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "74fe65ae-d0c9-4568-ba80-7b4071d6004f", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:21.123Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "7f0fee31-0ded-4656-b02c-28d0b9214b31", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:23.692Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "fa05f68f-a034-4c33-bdd7-32fc6d537119", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:469\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:56:32.366Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "828d7f1e-ecb3-4cf4-b500-dddbe6ae08a8", "latest": "2026-05-17 00:56:28 +0000 UTC", "lastBackup": "on-pitr-minio"} [mysql] 2026/05/17 00:57:13 packets.go:58 read tcp 10.237.26.51:52540->10.237.26.59:33062: read: connection reset by peer 2026-05-17T00:57:13.092Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "4a70c491-aa35-4b5a-a9b8-2a7192191849", "err": "failed to connect to pod pitr-pxc-0: invalid connection"} 2026-05-17T00:57:18.785Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "a8bca86f-4088-4534-958e-9bbdbae72c3d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:57:24.456Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d29dc602-3443-4ab2-b44d-7068df70197c", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:57:30.447Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "62100c2b-f9c6-4c06-84f4-ce04893b08b6", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:57:35.929Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "4f453739-f907-4994-a4e3-f1b2a94a8f9c", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:57:41.578Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "635a34b0-a1d2-40d9-95ad-2373e3a90f61", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:57:47.247Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "26e6f42a-61a1-4347-bfc5-d3d2d9418c1c", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:57:52.895Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "ddd91f75-4af6-4767-bd08-1c2bb64b0159", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:57:58.573Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "a36fa9de-9322-4f19-94a2-d92fa9ed5489", "primary name": "pitr-pxc-0.pitr-pxc.pitr-29966.svc.cluster.local"} 2026-05-17T00:58:01.940Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c3b25ea5-5d53-4334-a69a-5f0b76fbe38f"} 2026-05-17T00:58:07.553Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f1e83b33-6b96-4686-8559-650e914b67b6"} 2026-05-17T00:58:13.062Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "6168e53b-a356-49c3-a9f0-c6ca75d2ed91"} 2026-05-17T00:58:18.640Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "41b76e8e-cf64-4cff-8e27-26a306bdf2f4"} 2026-05-17T00:58:24.748Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0ea9934a-701b-4341-8c91-1243030eefba"} 2026-05-17T00:58:30.242Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "b9c24eb5-7a9d-4640-969a-8756dd8e2ede"} 2026-05-17T00:58:36.242Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "8f9895af-9f9b-49ad-b328-bc0638a4f900"} 2026-05-17T00:58:42.336Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f4d3cac6-46c1-43b4-8968-824012da4cfb"} 2026-05-17T00:58:47.836Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "884ef411-1c69-4672-bc5b-1a06869fc0b7"} 2026-05-17T00:58:53.658Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "bc7dcb1f-b30f-44be-bb0f-87bb9a39f1fb"} 2026-05-17T00:58:59.534Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "b5ca04b0-5e58-4d3b-a941-db2cfb67b65d"} 2026-05-17T00:59:05.553Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "dc4be7aa-1814-484a-8d0d-2f1844768f2f"} 2026-05-17T00:59:11.341Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "2c5f8799-2caf-4986-810c-147ee2946050"} 2026-05-17T00:59:17.468Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c3c45841-4348-4200-baab-5b3c569b3c04"} 2026-05-17T00:59:22.252Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "9a35ef4e-fe1c-43af-871b-abda7dadf969", "latest": "2026-05-17 00:59:17 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-05-17T00:59:23.159Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "9a35ef4e-fe1c-43af-871b-abda7dadf969"} 2026-05-17T00:59:28.944Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f96718db-8ff8-4ad0-87b5-f571a63eeaec"} 2026-05-17T00:59:35.045Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d020bd0f-61e5-4717-91de-8b407feeff5f"} 2026-05-17T00:59:41.341Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "2edf1331-8da1-403c-a3ec-f620f8eb5fae"} 2026-05-17T00:59:46.353Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "6017a66e-27f1-411d-b87d-00deabc7b153"} 2026-05-17T00:59:50.491Z INFO stopping cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "98a7acd4-c644-42aa-b020-0720a6a55f9d", "cluster": "pitr"} 2026-05-17T00:59:52.217Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "b68a27f8-072e-49d6-8291-e58397ceeef8"} 2026-05-17T00:59:52.418Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "b68a27f8-072e-49d6-8291-e58397ceeef8", "error": "exec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc: failed to execute command in pod: unable to upgrade connection: pods \"pitr-pitr-6959bd6dc5-vhqlc\" not found", "errorVerbose": "unable to upgrade connection: pods \"pitr-pitr-6959bd6dc5-vhqlc\" not found\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.CheckPITRErrors\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:73\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771\nexec binlog collector pod pitr-pitr-6959bd6dc5-vhqlc\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.CheckPITRErrors\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:78\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:221\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:478\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1771"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:494 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:437 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.24.1/pkg/internal/controller/controller.go:312 2026-05-17T00:59:52.420Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "17c20470-f049-4e1e-884d-501a46dc1fc8", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "0e68f460-9e35-45e0-a3fb-31d884297194", +  UID: "", -  ResourceVersion: "1778979478824895022", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:55:17 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:57:58 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &3, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-78bfbcdb4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-78bfbcdb4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-05-17T00:59:52.468Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "17c20470-f049-4e1e-884d-501a46dc1fc8", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "579a141a-8144-4723-8047-1f36f9842b15", +  UID: "", -  ResourceVersion: "1778979301836047023", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:40:55 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:55:01 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-6cc6944f96", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-6cc6944f96", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-05-17T00:59:52.561Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "17c20470-f049-4e1e-884d-501a46dc1fc8", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "579a141a-8144-4723-8047-1f36f9842b15", +  UID: "", -  ResourceVersion: "1778979301836047023", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:40:55 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:55:01 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-6cc6944f96", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-6cc6944f96", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-05-17T01:00:34.711Z INFO starting restore {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "dbfc1647-a00f-48c4-87c1-144d3a8c07c0", "cluster": "pitr", "backup": "on-pitr-minio"} 2026-05-17T01:00:34.775Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "6396c395-4b2e-44cf-bb9b-7ae052c8c514", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:00:39.805Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2cded95d-ce86-4519-acd4-be6d61bbdc43", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:00:44.826Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b72d1b0e-26b2-4119-9604-0eaa7d603647", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:00:50.108Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ad580390-07c8-4f6b-8fcb-0e44a9c5a1fd", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:00:55.123Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4e9277c3-ce72-4a6d-b488-d3989fcd5e60", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:00.243Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f9971a90-70e0-4c83-8b42-8cc49100e3b4", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:05.264Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1de8cefd-713f-4435-b8f8-1ebb088a2698", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-05-17T01:01:05.265Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1de8cefd-713f-4435-b8f8-1ebb088a2698", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-05-17T01:01:05.266Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1de8cefd-713f-4435-b8f8-1ebb088a2698", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-05-17T01:01:05.266Z INFO preparing cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1de8cefd-713f-4435-b8f8-1ebb088a2698", "cluster": "pitr"} 2026-05-17T01:01:05.344Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7e957116-d232-4575-a5e6-eee64882476c", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:10.362Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2d4cd13c-fee3-413b-a1ec-817f47d543b7", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:15.502Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b85dbc4f-765d-4005-87d6-a04d73b91f9c", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:20.519Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8eacb6de-8470-484d-8305-88a30294b897", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:25.535Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5698cafe-6fea-425a-9e5a-f6ae18e1e162", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:30.562Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "679312e2-16a9-4cf7-802e-f86850a56f94", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:35.584Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "600b043e-ac6b-464d-8637-9fa147603783", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:40.602Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "34e712a3-1c08-4c9c-92d8-e5b8d165f345", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:45.624Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4a3f6650-5be9-4a89-8218-b5402781a8f8", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:50.639Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a499db41-d082-4496-a94d-12525f9f9e82", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:01:55.654Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9f50ec42-b24a-4098-a2d5-9d972ef2b03b", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:02:00.715Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c6bf808a-8fbd-4c0f-9f3a-3bfd0d27889a", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:02:05.731Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c55e2c4d-3727-4f28-a25f-b37ecf851653", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:02:10.746Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f106a151-963d-4429-ae3f-4bd386bcab39", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:02:15.762Z INFO Waiting for prepare job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "361063f3-6230-4870-92e1-17194e6b73d5", "job": "prepare-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:02:20.945Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "eb6bdfc3-b896-4135-82c1-7ee228ef140b", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "0e68f460-9e35-45e0-a3fb-31d884297194", +  UID: "", -  ResourceVersion: "1778979633083103022", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:59:52 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 01:00:33 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-78bfbcdb4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-78bfbcdb4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-05-17T01:02:21.022Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "eb6bdfc3-b896-4135-82c1-7ee228ef140b", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "579a141a-8144-4723-8047-1f36f9842b15", +  UID: "", -  ResourceVersion: "1778979594325199023", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiI5MGQwNDg1Y2ZkZjc2OWIxZjZkOGVlMDYxY2JlNTIzZSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiOGY5ZGRjMDdjZTQ0MTI0ZDRkNDA3Y2VjMTk1NmFiODYifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiI5MGQwNDg1Y2ZkZjc2OWIxZjZkOGVlMDYxY2JlNTIzZSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiOGY5ZGRjMDdjZTQ0MTI0ZDRkNDA3Y2VjMTk1NmFiODYifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:59:52 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:59:54 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-6cc6944f96", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-6cc6944f96", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-05-17T01:02:21.090Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "eb6bdfc3-b896-4135-82c1-7ee228ef140b", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "579a141a-8144-4723-8047-1f36f9842b15", +  UID: "", -  ResourceVersion: "1778979594325199023", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiI5MGQwNDg1Y2ZkZjc2OWIxZjZkOGVlMDYxY2JlNTIzZSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiOGY5ZGRjMDdjZTQ0MTI0ZDRkNDA3Y2VjMTk1NmFiODYifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiI5MGQwNDg1Y2ZkZjc2OWIxZjZkOGVlMDYxY2JlNTIzZSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiOGY5ZGRjMDdjZTQ0MTI0ZDRkNDA3Y2VjMTk1NmFiODYifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0NzMtNmQzOTJiZWEiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:59:52 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 00:59:54 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-6cc6944f96", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-6cc6944f96", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-05-17T01:02:25.853Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8389b8e2-f308-4320-b5cd-6940bf0a5d34", "cluster": "pitr"} 2026-05-17T01:02:30.946Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4a738af8-4c4f-4f72-afe2-03f69920e167", "cluster": "pitr"} 2026-05-17T01:02:35.968Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "63f35d0d-5cbc-436a-9a99-43a69428dabf", "cluster": "pitr"} 2026-05-17T01:02:40.987Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8a56776a-ddc2-405b-8a28-65c26e857716", "cluster": "pitr"} 2026-05-17T01:02:46.009Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8b38bf79-14cc-492d-a1e3-211ad07fbbc4", "cluster": "pitr"} 2026-05-17T01:02:51.040Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b6f80189-5bf2-475d-9ab1-8c002b8340cd", "cluster": "pitr"} 2026-05-17T01:02:56.056Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fd7e0c24-a147-4120-ad08-e2c3b456c6ce", "cluster": "pitr"} 2026-05-17T01:03:01.073Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a30e343c-4add-4f7e-850c-7ad08c78af06", "cluster": "pitr"} 2026-05-17T01:03:03.067Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "44d1cd24-0f74-4a67-98d6-4f60badcdfdc"} 2026-05-17T01:03:03.132Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "44d1cd24-0f74-4a67-98d6-4f60badcdfdc", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T01:03:03.132Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "44d1cd24-0f74-4a67-98d6-4f60badcdfdc", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-05-17T01:03:06.090Z INFO point-in-time recovering {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7b1d4700-921a-48ed-9ded-fb3558d1ecd6", "cluster": "pitr"} 2026-05-17T01:03:06.154Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ffbf2a6d-fdae-43bb-9244-2bdb867df46d", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:03:08.302Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c42839c3-8086-4c08-ba2b-9697b5643311"} 2026-05-17T01:03:08.391Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c42839c3-8086-4c08-ba2b-9697b5643311", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T01:03:11.169Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "692fc7b7-9c99-4f39-b740-dd608f59af80", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-05-17T01:03:13.853Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f813d553-db1a-430a-82fd-e5daa8410fe3", "user": "xtrabackup"} 2026-05-17T01:03:13.861Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f813d553-db1a-430a-82fd-e5daa8410fe3"} 2026-05-17T01:03:13.897Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f813d553-db1a-430a-82fd-e5daa8410fe3", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T01:03:16.185Z INFO starting cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f6307696-6ffe-4b45-88fb-b4afee6b281f", "cluster": "pitr"} 2026-05-17T01:03:16.408Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f1f76fc8-4071-45c9-9829-1aca0582adb2", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "0e68f460-9e35-45e0-a3fb-31d884297194", +  UID: "", -  ResourceVersion: "1778979778629935022", +  ResourceVersion: "", -  Generation: 6, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 01:02:20 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 01:02:58 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &1, +  Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 6, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0, -  ReadyReplicas: 1, +  ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-78bfbcdb4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-78bfbcdb4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 1, +  AvailableReplicas: 0,    },   } 2026-05-17T01:03:16.462Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f1f76fc8-4071-45c9-9829-1aca0582adb2", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "579a141a-8144-4723-8047-1f36f9842b15", +  UID: "", -  ResourceVersion: "1778979741245727023", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 01:02:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 01:02:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-8bf9d5f4b", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-8bf9d5f4b", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-05-17T01:03:16.546Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f1f76fc8-4071-45c9-9829-1aca0582adb2", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-29966",    SelfLink: "", -  UID: "579a141a-8144-4723-8047-1f36f9842b15", +  UID: "", -  ResourceVersion: "1778979741245727023", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-05-17 00:40:55 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "606ced1d-87ab-4797-a5eb-e0e346e1b899", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 01:02:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-05-17 01:02:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "90d0485cfdf769b1f6d8ee061cbe523e", "percona.com/ssl-internal-hash": "8f9ddc07ce44124d4d407cec1956ab86"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-8bf9d5f4b", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-8bf9d5f4b", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-05-17T01:03:18.925Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "f1f76fc8-4071-45c9-9829-1aca0582adb2", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T01:03:21.270Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "18f2b61b-78c6-4667-ac0d-9bf5a149d2fd", "cluster": "pitr"} 2026-05-17T01:03:21.877Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "15cb4060-8f63-4025-b623-8800f36e2d34", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-29966 on 34.118.224.10:53: no such host"} 2026-05-17T01:03:26.298Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b3c1e6bc-61fc-4e4b-9629-e2a321f496be", "cluster": "pitr"} 2026-05-17T01:03:28.974Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "968ccffc-b6d1-4c31-90c1-1bf1d8e57eb5", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 10.237.24.69:6032: connect: connection refused"} 2026-05-17T01:03:31.314Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b7ef42d8-ea2e-42d7-9fd0-deea23441ca7", "cluster": "pitr"} 2026-05-17T01:03:36.341Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a7efe542-686f-4789-8a96-200885fc3919", "cluster": "pitr"} 2026-05-17T01:03:41.362Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e1bbe872-bef6-4eb9-8a63-2c09e02c1a40", "cluster": "pitr"} 2026-05-17T01:03:46.379Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "6d25aeba-60f5-459d-bd83-21e596ed5190", "cluster": "pitr"} 2026-05-17T01:03:51.398Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0876300f-bdbb-4988-949e-9f3f98f8e720", "cluster": "pitr"} 2026-05-17T01:03:56.417Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d833577c-411b-48aa-aa66-e180f3ee80ab", "cluster": "pitr"} 2026-05-17T01:04:01.436Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "edf3f6b7-24c6-48f8-9f1f-78bc11034633", "cluster": "pitr"} 2026-05-17T01:04:06.451Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "12dce97a-62cd-4313-90dc-d3017e0b8e02", "cluster": "pitr"} 2026-05-17T01:04:11.467Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "39f6d633-7117-41e1-bb29-bfdb2d9eebc8", "cluster": "pitr"} 2026-05-17T01:04:16.482Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cff6f5ab-838f-40e0-9b7d-0cda488d4a79", "cluster": "pitr"} 2026-05-17T01:04:21.503Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2fb4d1d4-8836-402a-b2d8-e11746857027", "cluster": "pitr"} 2026-05-17T01:04:26.518Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "17843810-db2a-4ec3-b63e-ce381e9b5514", "cluster": "pitr"} 2026-05-17T01:04:31.532Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "06985deb-c6f0-43a1-a215-efe6d5887837", "cluster": "pitr"} 2026-05-17T01:04:36.545Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d3c389f4-5775-486a-b9e7-ed9c8787d6b0", "cluster": "pitr"} 2026-05-17T01:04:41.561Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "76242af6-fe0b-4e3b-a2b5-c3c915f6c36e", "cluster": "pitr"} 2026-05-17T01:04:46.577Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "dac264cf-c56a-4bad-8611-bf564c9df6f0", "cluster": "pitr"} 2026-05-17T01:04:51.594Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cd04a067-b038-49da-886a-20e62786ddad", "cluster": "pitr"} 2026-05-17T01:04:56.609Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a551f404-8b83-4fb0-8602-4b7bf83cecaf", "cluster": "pitr"} 2026-05-17T01:05:01.625Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "278b96c2-79f6-475c-8c4b-d60d226b111e", "cluster": "pitr"} 2026-05-17T01:05:06.640Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "bfefb96b-2f4c-4548-95f6-a34ce8f5d665", "cluster": "pitr"} 2026-05-17T01:05:11.657Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "617ac74d-bd8d-4ce6-95f3-81b8b20484a9", "cluster": "pitr"} 2026-05-17T01:05:16.672Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1c7d8da8-2f7f-4416-9070-f587cd231d0f", "cluster": "pitr"} 2026-05-17T01:05:21.689Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fff95da9-e61c-4a0f-bd23-ca28f22f0ada", "cluster": "pitr"} 2026-05-17T01:05:26.705Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a3d16eb7-365c-4ab3-9d8a-45975dc4f07e", "cluster": "pitr"} 2026-05-17T01:05:31.720Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "76d97d34-c2c0-4620-88e5-df72f7e8865c", "cluster": "pitr"} 2026-05-17T01:05:36.737Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "633807a6-7783-4529-95ca-6ef9498a7620", "cluster": "pitr"} 2026-05-17T01:05:41.751Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "851a985a-b173-42b9-9ede-58ac7323dada", "cluster": "pitr"} 2026-05-17T01:05:46.774Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "be79da21-84a0-4c3f-b392-f250a167c0dc", "cluster": "pitr"} 2026-05-17T01:05:51.792Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "46ea44d8-74df-4043-846b-8538ba977b45", "cluster": "pitr"} 2026-05-17T01:05:56.809Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3b3209db-05cc-4ae9-8cb1-14fb9169e767", "cluster": "pitr"} 2026-05-17T01:06:02.005Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d4e4a694-0fd0-46ee-9beb-b5609238d0f0"} 2026-05-17T01:06:05.644Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "804e63c6-89de-4431-848f-b243b658cce7", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-05-17T01:06:07.299Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "804e63c6-89de-4431-848f-b243b658cce7"} 2026-05-17T01:06:12.696Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "9c2f3ac8-3a3f-40f8-ad42-a0d4d5c949f5"} 2026-05-17T01:06:17.357Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "cfa4c9cd-18bc-4848-a501-e1f81eff028d", "latest": "2026-05-17 01:04:03 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-05-17T01:06:18.498Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "cfa4c9cd-18bc-4848-a501-e1f81eff028d"} 2026-05-17T01:06:24.306Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "e60563b0-cde5-4600-bda1-e0ed8b63a5e5"} 2026-05-17T01:06:29.805Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "c9569cfa-c9d6-472e-ae9f-2dca4176694a"} 2026-05-17T01:06:35.702Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "9059f4fc-11f0-4a58-ae10-9d9a70cb7f5a"} 2026-05-17T01:06:41.895Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "cbe6a57f-20c4-42a7-b3ac-666ec3d95179"} 2026-05-17T01:06:47.693Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "0778c7d8-28de-40bc-aa36-b47ffcab6c5a"} 2026-05-17T01:06:53.700Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "b88234e9-ee3f-47a8-8990-03324e251bd8"} 2026-05-17T01:06:59.297Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "776132b9-d895-414f-9c1a-206f5a547e74"} 2026-05-17T01:07:05.001Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "9ed0be89-3c6b-44df-a47e-da607390084c"} 2026-05-17T01:07:11.189Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "48c391a6-c0a4-4ce5-a167-c283b1265645"} 2026-05-17T01:07:16.505Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "7064ebd8-e4e3-494e-9c66-91fc544b4dd7"} 2026-05-17T01:07:22.799Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "2bc9cfd9-6acb-4401-b375-3776b84c18bd"} 2026-05-17T01:07:28.506Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "6b36db0f-6cce-4c05-8ef2-db0c25c8c8de"} 2026-05-17T01:07:34.598Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "a574a7f1-8817-4803-97ea-596bdfbe8812"} 2026-05-17T01:07:40.494Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d9f8fa4d-5af7-4ee6-b42f-e697c22401e4"} 2026-05-17T01:07:46.197Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "6437ba14-4385-4727-898a-ed5201fdcc9a"} 2026-05-17T01:07:51.804Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "aee4e7e5-b02f-4afb-93c5-3f248a7d2f40"} 2026-05-17T01:07:57.901Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "9dbb7d9f-c887-4e05-8c69-459a40358d01"} 2026-05-17T01:08:03.496Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "7fa2d0ca-0572-4c77-abcc-b7eb05c8c423"} 2026-05-17T01:08:09.690Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "ad8f62a4-c004-4dba-8588-89b2d6e20a81"} 2026-05-17T01:08:15.502Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "dbff4066-f3bb-44c2-8472-ce3b8e0a2101"} 2026-05-17T01:08:21.294Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "d5406974-1f8b-490b-ab1d-c638431074a6"} 2026-05-17T01:08:27.398Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-29966"}, "namespace": "pitr-29966", "name": "pitr", "reconcileID": "fe2d238f-8b34-49b7-a8d2-2de3d74b41ae"} max retry count 120 reached. something went wrong with operator or kubernetes cluster