Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/logs/pitr-8-4.log Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 + GTID_PATTERN='[A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + main + create_infra pitr-31942 + local ns=pitr-31942 + '[' -n pxc-operator ']' + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + kubectl patch pxc -n pitr-19050 pitr --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/pitr patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.pjn0jYi7JP ++ mktemp + local LAST_ERR=/tmp/tmp.fFUWWDuQY9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pjn0jYi7JP perconaxtradbcluster.pxc.percona.com "pitr" deleted from pitr-19050 namespace + cat /tmp/tmp.fFUWWDuQY9 + rm /tmp/tmp.pjn0jYi7JP /tmp/tmp.fFUWWDuQY9 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.zzLW4qBBvH ++ mktemp + local LAST_ERR=/tmp/tmp.4jiSBs4fcA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zzLW4qBBvH perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio" deleted from pitr-19050 namespace perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio-2" deleted from pitr-19050 namespace perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio-3" deleted from pitr-19050 namespace + cat /tmp/tmp.4jiSBs4fcA + rm /tmp/tmp.zzLW4qBBvH /tmp/tmp.4jiSBs4fcA + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.SziY4PTdSn ++ mktemp + local LAST_ERR=/tmp/tmp.6FspVoJYZU + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SziY4PTdSn perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio" deleted from pitr-19050 namespace + cat /tmp/tmp.6FspVoJYZU + rm /tmp/tmp.SziY4PTdSn /tmp/tmp.6FspVoJYZU + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ awk '-F ' '{print $2}' ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator ++ mktemp + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.Zoj4y4mNpe ++ mktemp + local LAST_ERR=/tmp/tmp.gYovP02oeP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.Fb0C9Uw662 ++ mktemp + local LAST_ERR=/tmp/tmp.ioWAT2nWp7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Fb0C9Uw662 + cat /tmp/tmp.ioWAT2nWp7 + rm /tmp/tmp.Fb0C9Uw662 /tmp/tmp.ioWAT2nWp7 + return 0 namespace "cert-manager" deleted namespace "pitr-19050" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Zoj4y4mNpe namespace "pxc-operator" deleted + cat /tmp/tmp.gYovP02oeP + rm /tmp/tmp.Zoj4y4mNpe /tmp/tmp.gYovP02oeP + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.0IMXZtyxcU ++ mktemp + local LAST_ERR=/tmp/tmp.RngFUNNdle + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0IMXZtyxcU namespace/pxc-operator created + cat /tmp/tmp.RngFUNNdle + rm /tmp/tmp.0IMXZtyxcU /tmp/tmp.RngFUNNdle + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.28ILt5aqzb +++ mktemp ++ local LAST_ERR=/tmp/tmp.cnwWdw1vQf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.28ILt5aqzb ++ cat /tmp/tmp.cnwWdw1vQf ++ rm /tmp/tmp.28ILt5aqzb /tmp/tmp.cnwWdw1vQf ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster6 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.AsH2Fiz2wy ++ mktemp + local LAST_ERR=/tmp/tmp.AiO88ezz36 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster6 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.AsH2Fiz2wy Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster6" modified. + cat /tmp/tmp.AiO88ezz36 + rm /tmp/tmp.AsH2Fiz2wy /tmp/tmp.AiO88ezz36 + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.WLZ2Kmssc1 ++ mktemp + local LAST_ERR=/tmp/tmp.qkCHhS62n7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WLZ2Kmssc1 customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.qkCHhS62n7 + rm /tmp/tmp.WLZ2Kmssc1 /tmp/tmp.qkCHhS62n7 + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-rbac.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.sOymgMxGqn + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + local LAST_ERR=/tmp/tmp.BGJjj8CTur + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sOymgMxGqn clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.BGJjj8CTur + rm /tmp/tmp.sOymgMxGqn /tmp/tmp.BGJjj8CTur + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = ""' - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-operator.yaml + kubectl_bin apply -f - ++ mktemp + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + local LAST_OUT=/tmp/tmp.b5L85Y8nxV ++ mktemp + local LAST_ERR=/tmp/tmp.HzHEtm9Kb4 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.b5L85Y8nxV deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.HzHEtm9Kb4 + rm /tmp/tmp.b5L85Y8nxV /tmp/tmp.HzHEtm9Kb4 + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.C0XYtkCA7s ++ mktemp + local LAST_ERR=/tmp/tmp.xzy1iJ5Te8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.C0XYtkCA7s pod/percona-xtradb-cluster-operator-7cd7bb76f4-kwnbc condition met + cat /tmp/tmp.xzy1iJ5Te8 + rm /tmp/tmp.C0XYtkCA7s /tmp/tmp.xzy1iJ5Te8 + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ grep -c percona-xtradb-cluster-operator +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ head -1 ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.lsdQPl27op +++ mktemp ++ local LAST_ERR=/tmp/tmp.M0iBZj5pTM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lsdQPl27op ++ cat /tmp/tmp.M0iBZj5pTM ++ rm /tmp/tmp.lsdQPl27op /tmp/tmp.M0iBZj5pTM ++ return 0 + wait_pod percona-xtradb-cluster-operator-7cd7bb76f4-kwnbc 480 pxc-operator + local pod=percona-xtradb-cluster-operator-7cd7bb76f4-kwnbc + local max_retry=480 + local ns=pxc-operator ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo percona-xtradb-cluster-operator-7cd7bb76f4-kwnbc + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-7cd7bb76f4-kwnbc condition met waiting for pod/percona-xtradb-cluster-operator-7cd7bb76f4-kwnbc to become Ready.Ok + sleep 3 + create_namespace pitr-31942 + local namespace=pitr-31942 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pitr-31942' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pitr-31942 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pitr-31942 + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + kubectl_bin get ns + awk '{print$1}' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.dnt8ubxyia ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.teO8yhsGuM ++ mktemp + local LAST_ERR=/tmp/tmp.4hQgJJqn6y + local exit_status=0 + local LAST_ERR=/tmp/tmp.i4x3DkuI93 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-31942 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-31942 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.teO8yhsGuM + cat /tmp/tmp.i4x3DkuI93 + rm /tmp/tmp.teO8yhsGuM /tmp/tmp.i4x3DkuI93 + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-31942 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.dnt8ubxyia + cat /tmp/tmp.4hQgJJqn6y Error from server (NotFound): namespaces "pitr-31942" not found + rm /tmp/tmp.dnt8ubxyia /tmp/tmp.4hQgJJqn6y + return 1 + : + wait_for_delete namespace/pitr-31942 + local res=namespace/pitr-31942 + echo -n 'waiting for namespace/pitr-31942 to be deleted' waiting for namespace/pitr-31942 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pitr-31942" not found + desc 'create namespace pitr-31942' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pitr-31942 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pitr-31942 ++ mktemp + local LAST_OUT=/tmp/tmp.MlP0rPl12v ++ mktemp + local LAST_ERR=/tmp/tmp.CqalWx1bk0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pitr-31942 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.MlP0rPl12v namespace/pitr-31942 created + cat /tmp/tmp.CqalWx1bk0 + rm /tmp/tmp.MlP0rPl12v /tmp/tmp.CqalWx1bk0 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.uwOyVIdTrJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.XOyswFmCxH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uwOyVIdTrJ ++ cat /tmp/tmp.XOyswFmCxH ++ rm /tmp/tmp.uwOyVIdTrJ /tmp/tmp.XOyswFmCxH ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster6 --namespace=pitr-31942 ++ mktemp + local LAST_OUT=/tmp/tmp.LBqrj6C9Wj ++ mktemp + local LAST_ERR=/tmp/tmp.fVzElcj6I9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster6 --namespace=pitr-31942 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LBqrj6C9Wj Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster6" modified. + cat /tmp/tmp.fVzElcj6I9 + rm /tmp/tmp.LBqrj6C9Wj /tmp/tmp.fVzElcj6I9 + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.diRum3nOjs ++ mktemp + local LAST_ERR=/tmp/tmp.9HotvHb29W + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.diRum3nOjs secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.9HotvHb29W + rm /tmp/tmp.diRum3nOjs /tmp/tmp.9HotvHb29W + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.RqebOPZ0oN ++ mktemp + local LAST_ERR=/tmp/tmp.5aje9qs4Q8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RqebOPZ0oN namespace/cert-manager created + cat /tmp/tmp.5aje9qs4Q8 + rm /tmp/tmp.RqebOPZ0oN /tmp/tmp.5aje9qs4Q8 + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.oRDs2RKc8Q ++ mktemp + local LAST_ERR=/tmp/tmp.DwLbgFxgsI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oRDs2RKc8Q namespace/cert-manager labeled + cat /tmp/tmp.DwLbgFxgsI + rm /tmp/tmp.oRDs2RKc8Q /tmp/tmp.DwLbgFxgsI + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.EIpILNuKXG ++ mktemp + local LAST_ERR=/tmp/tmp.rlyFTGpZpo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EIpILNuKXG namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.rlyFTGpZpo Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.EIpILNuKXG /tmp/tmp.rlyFTGpZpo + return 0 + '[' '' == 4.10 ']' + sleep 70 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/issuer.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9J1OZya6vI ++ mktemp + local LAST_ERR=/tmp/tmp.o6QTA7mPLa + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/issuer.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9J1OZya6vI issuer.cert-manager.io/selfsigning-issuer created + cat /tmp/tmp.o6QTA7mPLa + rm /tmp/tmp.9J1OZya6vI /tmp/tmp.o6QTA7mPLa + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/cert.yml ++ mktemp + local LAST_OUT=/tmp/tmp.LxXvHHxukB ++ mktemp + local LAST_ERR=/tmp/tmp.8hrhZIfew9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/cert.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LxXvHHxukB certificate.cert-manager.io/tls-minio created + cat /tmp/tmp.8hrhZIfew9 Warning: spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. + rm /tmp/tmp.LxXvHHxukB /tmp/tmp.8hrhZIfew9 + return 0 + sleep_with_log 25 + local d=25 + log 'sleeping for 25 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:43:26+0000]' sleeping for 25 seconds [2026-04-29T16:43:26+0000] sleeping for 25 seconds + sleep 25 + start_minio tls-minio + deploy_helm pitr-31942 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=tls-minio + local endpoint=http://minio-service:9000 + minio_args=(--version $MINIO_VER --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set "users[0].accessKey=some-access-key" --set "users[0].secretKey=some-secret-key" --set "users[0].policy=consoleAdmin" --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G) + local minio_args + [[ -n tls-minio ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Wed Apr 29 16:43:57 2026 NAMESPACE: pitr-31942 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.pitr-31942.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace pitr-31942 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace pitr-31942 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace pitr-31942 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace pitr-31942 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xb7Y0bR8qU +++ mktemp ++ local LAST_ERR=/tmp/tmp.dXMSaHtdHP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xb7Y0bR8qU ++ cat /tmp/tmp.dXMSaHtdHP ++ rm /tmp/tmp.Xb7Y0bR8qU /tmp/tmp.dXMSaHtdHP ++ return 0 + MINIO_POD=minio-service-757b74c96c-8xdn6 + wait_pod minio-service-757b74c96c-8xdn6 + local pod=minio-service-757b74c96c-8xdn6 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo minio-service-757b74c96c-8xdn6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/minio-service-757b74c96c-8xdn6 condition met waiting for pod/minio-service-757b74c96c-8xdn6 to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.V14pFoGMtE ++ mktemp + local LAST_ERR=/tmp/tmp.wxpTqfGm2p + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.V14pFoGMtE /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from pitr-31942 namespace + cat /tmp/tmp.wxpTqfGm2p + rm /tmp/tmp.V14pFoGMtE /tmp/tmp.wxpTqfGm2p + return 0 + cluster=pitr + spinup_pxc pitr /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local cluster=pitr + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.Roc8YlUwjw ++ mktemp + local LAST_ERR=/tmp/tmp.HsvSM4Cfhi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Roc8YlUwjw secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.HsvSM4Cfhi + rm /tmp/tmp.Roc8YlUwjw /tmp/tmp.HsvSM4Cfhi + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-31942~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + local LAST_OUT=/tmp/tmp.pj2gk8azcU + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ mktemp + local LAST_ERR=/tmp/tmp.64iPzysVGO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pj2gk8azcU deployment.apps/pxc-client created + cat /tmp/tmp.64iPzysVGO + rm /tmp/tmp.pj2gk8azcU /tmp/tmp.64iPzysVGO + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml '' + kubectl_bin apply -f - + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local pvc_name= + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-31942~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + local LAST_OUT=/tmp/tmp.zM8vnl9VUz + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' ++ mktemp + local LAST_ERR=/tmp/tmp.Luts8DkCtB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zM8vnl9VUz perconaxtradbcluster.pxc.percona.com/pitr created + cat /tmp/tmp.Luts8DkCtB + rm /tmp/tmp.zM8vnl9VUz /tmp/tmp.Luts8DkCtB + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lcBIsIKoIr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.6LMvvQbOSK +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.lcBIsIKoIr +++ cat /tmp/tmp.6LMvvQbOSK +++ rm /tmp/tmp.lcBIsIKoIr /tmp/tmp.6LMvvQbOSK +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.QlNinqoHpv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.kbFGWDOM94 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.QlNinqoHpv +++ cat /tmp/tmp.kbFGWDOM94 +++ rm /tmp/tmp.QlNinqoHpv /tmp/tmp.kbFGWDOM94 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-31942 ++ mktemp + local LAST_OUT=/tmp/tmp.YqO7gC6fUB ++ mktemp + local LAST_ERR=/tmp/tmp.E2miBd1uFl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-31942 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-31942 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-31942 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.YqO7gC6fUB + cat /tmp/tmp.E2miBd1uFl error: no matching resources found + rm /tmp/tmp.YqO7gC6fUB /tmp/tmp.E2miBd1uFl + return 1 + true + wait_for_running pitr-proxysql 1 + local name=pitr-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ echo pitr-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/pitr-proxysql-0 condition met waiting for pod/pitr-proxysql-0 to become Ready.Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-0 + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-1 + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc pitr -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ base64 --decode ++ local LAST_OUT=/tmp/tmp.Iaibdg3Exk +++ mktemp ++ local LAST_ERR=/tmp/tmp.Alm9CHUxPx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Iaibdg3Exk ++ cat /tmp/tmp.Alm9CHUxPx ++ rm /tmp/tmp.Iaibdg3Exk /tmp/tmp.Alm9CHUxPx ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZYKoZWXBaN +++ mktemp ++ local LAST_ERR=/tmp/tmp.aL4XzzKJkI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZYKoZWXBaN ++ cat /tmp/tmp.aL4XzzKJkI ++ rm /tmp/tmp.ZYKoZWXBaN /tmp/tmp.aL4XzzKJkI ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-6jk4n + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OXeEqkxg6c +++ mktemp ++ local LAST_ERR=/tmp/tmp.qiQYH4eWc9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OXeEqkxg6c ++ cat /tmp/tmp.qiQYH4eWc9 ++ rm /tmp/tmp.OXeEqkxg6c /tmp/tmp.qiQYH4eWc9 ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.py4HXAsexD +++ mktemp ++ local LAST_ERR=/tmp/tmp.aJDH4KvEWr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.py4HXAsexD ++ cat /tmp/tmp.aJDH4KvEWr ++ rm /tmp/tmp.py4HXAsexD /tmp/tmp.aJDH4KvEWr ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.4FyEIG03fJ/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UhHxjhOMqc +++ mktemp ++ local LAST_ERR=/tmp/tmp.AZhVTpvkS3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UhHxjhOMqc ++ cat /tmp/tmp.AZhVTpvkS3 ++ rm /tmp/tmp.UhHxjhOMqc /tmp/tmp.AZhVTpvkS3 ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.4FyEIG03fJ/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VNj02GWHdt +++ mktemp ++ local LAST_ERR=/tmp/tmp.pOAvFdHdNM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VNj02GWHdt ++ cat /tmp/tmp.pOAvFdHdNM ++ rm /tmp/tmp.VNj02GWHdt /tmp/tmp.pOAvFdHdNM ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.4FyEIG03fJ/select-1.sql + is_keyring_plugin_in_use pitr + local cluster=pitr + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + kubectl exec pitr-pxc-0 -c pxc -- ls /var/lib/mysql/mysqld.my + return 1 ++ kubectl_bin get pods --selector=app.kubernetes.io/component=pitr -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nMQigVOLaZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.DUzpqspIo1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/component=pitr -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nMQigVOLaZ ++ cat /tmp/tmp.DUzpqspIo1 ++ rm /tmp/tmp.nMQigVOLaZ /tmp/tmp.DUzpqspIo1 ++ return 0 + pitr_pod=pitr-pitr-798584cd4f-jzrm4 + wait_pod pitr-pitr-798584cd4f-jzrm4 + local pod=pitr-pitr-798584cd4f-jzrm4 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-pitr-798584cd4f-jzrm4 + local container= + set +o xtrace pod/pitr-pitr-798584cd4f-jzrm4 condition met waiting for pod/pitr-pitr-798584cd4f-jzrm4 to become ReadyDefaulted container "pitr" out of: pitr, pitr-init (init) .Ok ++ kubectl_bin logs -l app.kubernetes.io/component=pitr --tail=-1 ++ egrep -c 'already exists|is already installed' + [[ 1 == \0 ]] + run_backup pitr on-pitr-minio + local cluster=pitr + local backup=on-pitr-minio + local name= + '[' -z ']' + name=on-pitr-minio + log 'run backup pxc-backup/on-pitr-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:51:04+0000]' run backup pxc-backup/on-pitr-minio [2026-04-29T16:51:04+0000] run backup pxc-backup/on-pitr-minio + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/on-pitr-minio.yml + kubectl_bin apply -f - + yq eval '.metadata.name="on-pitr-minio"' ++ mktemp + local LAST_OUT=/tmp/tmp.P4CAUS3Ekh ++ mktemp + local LAST_ERR=/tmp/tmp.0tXCRjmddC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.P4CAUS3Ekh perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio created + cat /tmp/tmp.0tXCRjmddC + rm /tmp/tmp.P4CAUS3Ekh /tmp/tmp.0tXCRjmddC + return 0 + wait_backup on-pitr-minio + local backup=on-pitr-minio + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio to reach Succeeded state....................Succeeded + write_test_data pitr + local cluster=pitr + local config= + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0iASTyLf6n ++++ mktemp +++ local LAST_ERR=/tmp/tmp.q13q0KRph9 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.0iASTyLf6n +++ cat /tmp/tmp.q13q0KRph9 +++ rm /tmp/tmp.0iASTyLf6n /tmp/tmp.q13q0KRph9 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3rPutHkPIk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.G1eF0lciES +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.3rPutHkPIk +++ cat /tmp/tmp.G1eF0lciES +++ rm /tmp/tmp.3rPutHkPIk /tmp/tmp.G1eF0lciES +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + desc 'write test data' + set +o xtrace ----------------------------------------------------------------------------------- write test data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.keM6HFzb9y +++ mktemp ++ local LAST_ERR=/tmp/tmp.m955ipBnnX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.keM6HFzb9y ++ cat /tmp/tmp.m955ipBnnX ++ rm /tmp/tmp.keM6HFzb9y /tmp/tmp.m955ipBnnX ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-56fd5498cd-6jk4n + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' '-h pitr-proxysql -uroot -proot_password' + local 'command=INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zlo2ivKE71 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yQCPTlQiLH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Zlo2ivKE71 ++ cat /tmp/tmp.yQCPTlQiLH ++ rm /tmp/tmp.Zlo2ivKE71 /tmp/tmp.yQCPTlQiLH ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep_with_log 30 + local d=30 + log 'sleeping for 30 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:52:04+0000]' sleeping for 30 seconds [2026-04-29T16:52:04+0000] sleeping for 30 seconds + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9IZ0O3GB3Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.vk61qwL0ay ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9IZ0O3GB3Q ++ cat /tmp/tmp.vk61qwL0ay ++ rm /tmp/tmp.9IZ0O3GB3Q /tmp/tmp.vk61qwL0ay ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-56fd5498cd-6jk4n ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.4FyEIG03fJ/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2wZQaUf8v7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.XpMOkzwwC5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2wZQaUf8v7 ++ cat /tmp/tmp.XpMOkzwwC5 ++ rm /tmp/tmp.2wZQaUf8v7 /tmp/tmp.XpMOkzwwC5 ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.4FyEIG03fJ/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZYGzF7TNvf +++ mktemp ++ local LAST_ERR=/tmp/tmp.0sdL4zlGdo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZYGzF7TNvf ++ cat /tmp/tmp.0sdL4zlGdo ++ rm /tmp/tmp.ZYGzF7TNvf /tmp/tmp.0sdL4zlGdo ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.4FyEIG03fJ/select-2.sql ++ is_keyring_plugin_in_use pitr ++ local cluster=pitr ++ [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] ++ kubectl exec pitr-pxc-0 -c pxc -- ls /var/lib/mysql/mysqld.my ++ return 1 + '[' '' ']' + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ echo -n pass1 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczE= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczE= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.XNlsUjTwnQ ++ mktemp + local LAST_ERR=/tmp/tmp.zRwvIOUcMF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.XNlsUjTwnQ secret/my-cluster-secrets patched + cat /tmp/tmp.zRwvIOUcMF + rm /tmp/tmp.XNlsUjTwnQ /tmp/tmp.zRwvIOUcMF + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0oTDKvn7JL +++ mktemp ++ local LAST_ERR=/tmp/tmp.ocvTSotW1Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0oTDKvn7JL ++ cat /tmp/tmp.ocvTSotW1Z ++ rm /tmp/tmp.0oTDKvn7JL /tmp/tmp.ocvTSotW1Z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PtWww1JpJL +++ mktemp ++ local LAST_ERR=/tmp/tmp.ios6AL5J2l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PtWww1JpJL ++ cat /tmp/tmp.ios6AL5J2l ++ rm /tmp/tmp.PtWww1JpJL /tmp/tmp.ios6AL5J2l ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LqbuluflyY +++ mktemp ++ local LAST_ERR=/tmp/tmp.N7yB4XmBph ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LqbuluflyY ++ cat /tmp/tmp.N7yB4XmBph ++ rm /tmp/tmp.LqbuluflyY /tmp/tmp.N7yB4XmBph ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U25DD3YuXr +++ mktemp ++ local LAST_ERR=/tmp/tmp.kYsxXP9iQR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.U25DD3YuXr ++ cat /tmp/tmp.kYsxXP9iQR ++ rm /tmp/tmp.U25DD3YuXr /tmp/tmp.kYsxXP9iQR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UPabVcNHbM +++ mktemp ++ local LAST_ERR=/tmp/tmp.PqAt2E2L89 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UPabVcNHbM ++ cat /tmp/tmp.PqAt2E2L89 ++ rm /tmp/tmp.UPabVcNHbM /tmp/tmp.PqAt2E2L89 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1bFjlTU9lj +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wu3CA5V5Jf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1bFjlTU9lj ++ cat /tmp/tmp.Wu3CA5V5Jf ++ rm /tmp/tmp.1bFjlTU9lj /tmp/tmp.Wu3CA5V5Jf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OWpemqXo0R +++ mktemp ++ local LAST_ERR=/tmp/tmp.XtxUiBDrPQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OWpemqXo0R ++ cat /tmp/tmp.XtxUiBDrPQ ++ rm /tmp/tmp.OWpemqXo0R /tmp/tmp.XtxUiBDrPQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XtRH9kIVEn +++ mktemp ++ local LAST_ERR=/tmp/tmp.QDE0hxaCrO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XtRH9kIVEn ++ cat /tmp/tmp.QDE0hxaCrO ++ rm /tmp/tmp.XtRH9kIVEn /tmp/tmp.QDE0hxaCrO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.obaFH1ehPE +++ mktemp ++ local LAST_ERR=/tmp/tmp.bs29LomjOY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.obaFH1ehPE ++ cat /tmp/tmp.bs29LomjOY ++ rm /tmp/tmp.obaFH1ehPE /tmp/tmp.bs29LomjOY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XyVWHE9bIM +++ mktemp ++ local LAST_ERR=/tmp/tmp.emISOSOxIe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XyVWHE9bIM ++ cat /tmp/tmp.emISOSOxIe ++ rm /tmp/tmp.XyVWHE9bIM /tmp/tmp.emISOSOxIe ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gpA9ZsGTjC +++ mktemp ++ local LAST_ERR=/tmp/tmp.Galvvb1029 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gpA9ZsGTjC ++ cat /tmp/tmp.Galvvb1029 ++ rm /tmp/tmp.gpA9ZsGTjC /tmp/tmp.Galvvb1029 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2p36lDLMZ6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.uKjJh8FrmD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2p36lDLMZ6 ++ cat /tmp/tmp.uKjJh8FrmD ++ rm /tmp/tmp.2p36lDLMZ6 /tmp/tmp.uKjJh8FrmD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sT5O3LZVxO +++ mktemp ++ local LAST_ERR=/tmp/tmp.DxA9se2w0I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sT5O3LZVxO ++ cat /tmp/tmp.DxA9se2w0I ++ rm /tmp/tmp.sT5O3LZVxO /tmp/tmp.DxA9se2w0I ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qrpcCWPoRm +++ mktemp ++ local LAST_ERR=/tmp/tmp.2EtZ485OFj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qrpcCWPoRm ++ cat /tmp/tmp.2EtZ485OFj ++ rm /tmp/tmp.qrpcCWPoRm /tmp/tmp.2EtZ485OFj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mjqWY6pVVB +++ mktemp ++ local LAST_ERR=/tmp/tmp.CQfkTk6c6g ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mjqWY6pVVB ++ cat /tmp/tmp.CQfkTk6c6g ++ rm /tmp/tmp.mjqWY6pVVB /tmp/tmp.CQfkTk6c6g ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k6LPcRRwgk +++ mktemp ++ local LAST_ERR=/tmp/tmp.nevQY8yDOx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k6LPcRRwgk ++ cat /tmp/tmp.nevQY8yDOx ++ rm /tmp/tmp.k6LPcRRwgk /tmp/tmp.nevQY8yDOx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sL6azcVEFW +++ mktemp ++ local LAST_ERR=/tmp/tmp.btY9E0Kon4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sL6azcVEFW ++ cat /tmp/tmp.btY9E0Kon4 ++ rm /tmp/tmp.sL6azcVEFW /tmp/tmp.btY9E0Kon4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Lbg4Lrc1bK +++ mktemp ++ local LAST_ERR=/tmp/tmp.DGLyB4dRhY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Lbg4Lrc1bK ++ cat /tmp/tmp.DGLyB4dRhY ++ rm /tmp/tmp.Lbg4Lrc1bK /tmp/tmp.DGLyB4dRhY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PL9yeqfVyK +++ mktemp ++ local LAST_ERR=/tmp/tmp.HeqlMKzzBS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PL9yeqfVyK ++ cat /tmp/tmp.HeqlMKzzBS ++ rm /tmp/tmp.PL9yeqfVyK /tmp/tmp.HeqlMKzzBS ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PzWr01NjmW +++ mktemp ++ local LAST_ERR=/tmp/tmp.JeXvV6XeMM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PzWr01NjmW ++ cat /tmp/tmp.JeXvV6XeMM ++ rm /tmp/tmp.PzWr01NjmW /tmp/tmp.JeXvV6XeMM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FnYIdifwPX +++ mktemp ++ local LAST_ERR=/tmp/tmp.iF1F9uFkh5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FnYIdifwPX ++ cat /tmp/tmp.iF1F9uFkh5 ++ rm /tmp/tmp.FnYIdifwPX /tmp/tmp.iF1F9uFkh5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MkDBvbqsJw +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZTw2YWhg1b ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MkDBvbqsJw ++ cat /tmp/tmp.ZTw2YWhg1b ++ rm /tmp/tmp.MkDBvbqsJw /tmp/tmp.ZTw2YWhg1b ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KnCUV8OAIE +++ mktemp ++ local LAST_ERR=/tmp/tmp.2tjw14GX4W ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KnCUV8OAIE ++ cat /tmp/tmp.2tjw14GX4W ++ rm /tmp/tmp.KnCUV8OAIE /tmp/tmp.2tjw14GX4W ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GUdkBnucpI +++ mktemp ++ local LAST_ERR=/tmp/tmp.AlbLn0ouOl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GUdkBnucpI ++ cat /tmp/tmp.AlbLn0ouOl ++ rm /tmp/tmp.GUdkBnucpI /tmp/tmp.AlbLn0ouOl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nC4jSF4dlX +++ mktemp ++ local LAST_ERR=/tmp/tmp.9o44Pqrb5P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nC4jSF4dlX ++ cat /tmp/tmp.9o44Pqrb5P ++ rm /tmp/tmp.nC4jSF4dlX /tmp/tmp.9o44Pqrb5P ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mpO5vBolkX +++ mktemp ++ local LAST_ERR=/tmp/tmp.X3cY5KZ5Yl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mpO5vBolkX ++ cat /tmp/tmp.X3cY5KZ5Yl ++ rm /tmp/tmp.mpO5vBolkX /tmp/tmp.X3cY5KZ5Yl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tnGOgCYSoQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.A7gc375BBS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tnGOgCYSoQ ++ cat /tmp/tmp.A7gc375BBS ++ rm /tmp/tmp.tnGOgCYSoQ /tmp/tmp.A7gc375BBS ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cpP14JNffx +++ mktemp ++ local LAST_ERR=/tmp/tmp.bKXGgngrzg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cpP14JNffx ++ cat /tmp/tmp.bKXGgngrzg ++ rm /tmp/tmp.cpP14JNffx /tmp/tmp.bKXGgngrzg ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.4O8845SZpD ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.fSMq0f0Gwf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.4O8845SZpD +++++ cat /tmp/tmp.fSMq0f0Gwf +++++ rm /tmp/tmp.4O8845SZpD /tmp/tmp.fSMq0f0Gwf +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.SBBxYUgHAd ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.OJyd2vzC5e +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.SBBxYUgHAd +++++ cat /tmp/tmp.OJyd2vzC5e +++++ rm /tmp/tmp.SBBxYUgHAd /tmp/tmp.OJyd2vzC5e +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xutc4DpXap +++ mktemp ++ local LAST_ERR=/tmp/tmp.p7iTiPGdxK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xutc4DpXap ++ cat /tmp/tmp.p7iTiPGdxK ++ rm /tmp/tmp.Xutc4DpXap /tmp/tmp.p7iTiPGdxK ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ echo -n pass2 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczI= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.xit6mSwgQu ++ mktemp + local LAST_ERR=/tmp/tmp.BS69piM6vi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xit6mSwgQu secret/my-cluster-secrets patched + cat /tmp/tmp.BS69piM6vi + rm /tmp/tmp.xit6mSwgQu /tmp/tmp.BS69piM6vi + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DfJkfreXpm +++ mktemp ++ local LAST_ERR=/tmp/tmp.3wxyRVfkPC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DfJkfreXpm ++ cat /tmp/tmp.3wxyRVfkPC ++ rm /tmp/tmp.DfJkfreXpm /tmp/tmp.3wxyRVfkPC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ahXUw09wEP +++ mktemp ++ local LAST_ERR=/tmp/tmp.jFBmfl4W41 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ahXUw09wEP ++ cat /tmp/tmp.jFBmfl4W41 ++ rm /tmp/tmp.ahXUw09wEP /tmp/tmp.jFBmfl4W41 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MwejazzvLM +++ mktemp ++ local LAST_ERR=/tmp/tmp.ATKYPw1Svn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MwejazzvLM ++ cat /tmp/tmp.ATKYPw1Svn ++ rm /tmp/tmp.MwejazzvLM /tmp/tmp.ATKYPw1Svn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MkCiPlPdMf +++ mktemp ++ local LAST_ERR=/tmp/tmp.pbib2Ee35y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MkCiPlPdMf ++ cat /tmp/tmp.pbib2Ee35y ++ rm /tmp/tmp.MkCiPlPdMf /tmp/tmp.pbib2Ee35y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3zEMlFgDxn +++ mktemp ++ local LAST_ERR=/tmp/tmp.9Qg72q0I7b ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3zEMlFgDxn ++ cat /tmp/tmp.9Qg72q0I7b ++ rm /tmp/tmp.3zEMlFgDxn /tmp/tmp.9Qg72q0I7b ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ObxCouNeRG +++ mktemp ++ local LAST_ERR=/tmp/tmp.lCAqy5XLk0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ObxCouNeRG ++ cat /tmp/tmp.lCAqy5XLk0 ++ rm /tmp/tmp.ObxCouNeRG /tmp/tmp.lCAqy5XLk0 ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cQH1bRHS35 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9xWBqtl6vg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cQH1bRHS35 ++ cat /tmp/tmp.9xWBqtl6vg ++ rm /tmp/tmp.cQH1bRHS35 /tmp/tmp.9xWBqtl6vg ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CcInFvbwVd +++ mktemp ++ local LAST_ERR=/tmp/tmp.tK3oh0TTzQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CcInFvbwVd ++ cat /tmp/tmp.tK3oh0TTzQ ++ rm /tmp/tmp.CcInFvbwVd /tmp/tmp.tK3oh0TTzQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1vfxrkRDtc +++ mktemp ++ local LAST_ERR=/tmp/tmp.zZNjKNcllt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1vfxrkRDtc ++ cat /tmp/tmp.zZNjKNcllt ++ rm /tmp/tmp.1vfxrkRDtc /tmp/tmp.zZNjKNcllt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sqWVyGhPkX +++ mktemp ++ local LAST_ERR=/tmp/tmp.piAVMfgGXf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sqWVyGhPkX ++ cat /tmp/tmp.piAVMfgGXf ++ rm /tmp/tmp.sqWVyGhPkX /tmp/tmp.piAVMfgGXf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pR2jeBzXVU +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y1Spe3W7fG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pR2jeBzXVU ++ cat /tmp/tmp.Y1Spe3W7fG ++ rm /tmp/tmp.pR2jeBzXVU /tmp/tmp.Y1Spe3W7fG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ryIpqqslMr +++ mktemp ++ local LAST_ERR=/tmp/tmp.B2IlAdVpK6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ryIpqqslMr ++ cat /tmp/tmp.B2IlAdVpK6 ++ rm /tmp/tmp.ryIpqqslMr /tmp/tmp.B2IlAdVpK6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GYoarSMMgf +++ mktemp ++ local LAST_ERR=/tmp/tmp.RiVEnqT0Jj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GYoarSMMgf ++ cat /tmp/tmp.RiVEnqT0Jj ++ rm /tmp/tmp.GYoarSMMgf /tmp/tmp.RiVEnqT0Jj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.60zW8vChd0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.II7HaiglDf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.60zW8vChd0 ++ cat /tmp/tmp.II7HaiglDf ++ rm /tmp/tmp.60zW8vChd0 /tmp/tmp.II7HaiglDf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UR3FQegyq2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.uXpo866CXY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UR3FQegyq2 ++ cat /tmp/tmp.uXpo866CXY ++ rm /tmp/tmp.UR3FQegyq2 /tmp/tmp.uXpo866CXY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EPy3PgcwXV +++ mktemp ++ local LAST_ERR=/tmp/tmp.xoDYUUN16o ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EPy3PgcwXV ++ cat /tmp/tmp.xoDYUUN16o ++ rm /tmp/tmp.EPy3PgcwXV /tmp/tmp.xoDYUUN16o ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Scxi6EvojJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qb0K9qozhd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Scxi6EvojJ ++ cat /tmp/tmp.Qb0K9qozhd ++ rm /tmp/tmp.Scxi6EvojJ /tmp/tmp.Qb0K9qozhd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h6h3DKwW3D +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qxst4eGdXg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h6h3DKwW3D ++ cat /tmp/tmp.Qxst4eGdXg ++ rm /tmp/tmp.h6h3DKwW3D /tmp/tmp.Qxst4eGdXg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9qfibEuHlG +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dl54JVRcRn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9qfibEuHlG ++ cat /tmp/tmp.Dl54JVRcRn ++ rm /tmp/tmp.9qfibEuHlG /tmp/tmp.Dl54JVRcRn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RsE0D4c2SW +++ mktemp ++ local LAST_ERR=/tmp/tmp.FMPhR3f0cG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RsE0D4c2SW ++ cat /tmp/tmp.FMPhR3f0cG ++ rm /tmp/tmp.RsE0D4c2SW /tmp/tmp.FMPhR3f0cG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2a3sHL4dmD +++ mktemp ++ local LAST_ERR=/tmp/tmp.C4JlvYyecS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2a3sHL4dmD ++ cat /tmp/tmp.C4JlvYyecS ++ rm /tmp/tmp.2a3sHL4dmD /tmp/tmp.C4JlvYyecS ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mnqCbLk6H0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q9OwBRBDRI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mnqCbLk6H0 ++ cat /tmp/tmp.Q9OwBRBDRI ++ rm /tmp/tmp.mnqCbLk6H0 /tmp/tmp.Q9OwBRBDRI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aStgLYj30i +++ mktemp ++ local LAST_ERR=/tmp/tmp.7ov06Lrji6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aStgLYj30i ++ cat /tmp/tmp.7ov06Lrji6 ++ rm /tmp/tmp.aStgLYj30i /tmp/tmp.7ov06Lrji6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SySXzSi0FA +++ mktemp ++ local LAST_ERR=/tmp/tmp.kneeCHKH8k ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SySXzSi0FA ++ cat /tmp/tmp.kneeCHKH8k ++ rm /tmp/tmp.SySXzSi0FA /tmp/tmp.kneeCHKH8k ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yYLxtKSK7l +++ mktemp ++ local LAST_ERR=/tmp/tmp.MebSkOii9Q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yYLxtKSK7l ++ cat /tmp/tmp.MebSkOii9Q ++ rm /tmp/tmp.yYLxtKSK7l /tmp/tmp.MebSkOii9Q ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t5eFkkjYDo +++ mktemp ++ local LAST_ERR=/tmp/tmp.hwfI9APaDO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.t5eFkkjYDo ++ cat /tmp/tmp.hwfI9APaDO ++ rm /tmp/tmp.t5eFkkjYDo /tmp/tmp.hwfI9APaDO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WnbMy5WzJl +++ mktemp ++ local LAST_ERR=/tmp/tmp.aftSI20Sqk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WnbMy5WzJl ++ cat /tmp/tmp.aftSI20Sqk ++ rm /tmp/tmp.WnbMy5WzJl /tmp/tmp.aftSI20Sqk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 26 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K0MUnm1oqa +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q35haktnwg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.K0MUnm1oqa ++ cat /tmp/tmp.Q35haktnwg ++ rm /tmp/tmp.K0MUnm1oqa /tmp/tmp.Q35haktnwg ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RJkHmEiA9d +++ mktemp ++ local LAST_ERR=/tmp/tmp.ka9OiuhmsZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RJkHmEiA9d ++ cat /tmp/tmp.ka9OiuhmsZ ++ rm /tmp/tmp.RJkHmEiA9d /tmp/tmp.ka9OiuhmsZ ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.E77GH18DH4 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.4ZHlxSl46o +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.E77GH18DH4 +++++ cat /tmp/tmp.4ZHlxSl46o +++++ rm /tmp/tmp.E77GH18DH4 /tmp/tmp.4ZHlxSl46o +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.xx3z1wxMx4 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Mlw2qPcu3E +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.xx3z1wxMx4 +++++ cat /tmp/tmp.Mlw2qPcu3E +++++ rm /tmp/tmp.xx3z1wxMx4 /tmp/tmp.Mlw2qPcu3E +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XQWbrIvTIq +++ mktemp ++ local LAST_ERR=/tmp/tmp.LaZOAp2yRf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XQWbrIvTIq ++ cat /tmp/tmp.LaZOAp2yRf ++ rm /tmp/tmp.XQWbrIvTIq /tmp/tmp.LaZOAp2yRf ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ echo -n pass3 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczM= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczM= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.aohQaJV8aC ++ mktemp + local LAST_ERR=/tmp/tmp.U08vZTTcXo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.aohQaJV8aC secret/my-cluster-secrets patched + cat /tmp/tmp.U08vZTTcXo + rm /tmp/tmp.aohQaJV8aC /tmp/tmp.U08vZTTcXo + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2weQ9hy7MR +++ mktemp ++ local LAST_ERR=/tmp/tmp.Am4xRqtVHr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2weQ9hy7MR ++ cat /tmp/tmp.Am4xRqtVHr ++ rm /tmp/tmp.2weQ9hy7MR /tmp/tmp.Am4xRqtVHr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y1KyvyW5dp +++ mktemp ++ local LAST_ERR=/tmp/tmp.ryfZIQVyWy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y1KyvyW5dp ++ cat /tmp/tmp.ryfZIQVyWy ++ rm /tmp/tmp.Y1KyvyW5dp /tmp/tmp.ryfZIQVyWy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.35FB7SF94o +++ mktemp ++ local LAST_ERR=/tmp/tmp.INXSNYLUAv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.35FB7SF94o ++ cat /tmp/tmp.INXSNYLUAv ++ rm /tmp/tmp.35FB7SF94o /tmp/tmp.INXSNYLUAv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P23wYAAE1F +++ mktemp ++ local LAST_ERR=/tmp/tmp.O22U5mufte ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.P23wYAAE1F ++ cat /tmp/tmp.O22U5mufte ++ rm /tmp/tmp.P23wYAAE1F /tmp/tmp.O22U5mufte ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.omsBbbxiYw +++ mktemp ++ local LAST_ERR=/tmp/tmp.2FRNJcENiv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.omsBbbxiYw ++ cat /tmp/tmp.2FRNJcENiv ++ rm /tmp/tmp.omsBbbxiYw /tmp/tmp.2FRNJcENiv ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZWB5rvYh0b +++ mktemp ++ local LAST_ERR=/tmp/tmp.5uy5hqpFUZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZWB5rvYh0b ++ cat /tmp/tmp.5uy5hqpFUZ ++ rm /tmp/tmp.ZWB5rvYh0b /tmp/tmp.5uy5hqpFUZ ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZPwu62rD7o +++ mktemp ++ local LAST_ERR=/tmp/tmp.VkffeZHccK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZPwu62rD7o ++ cat /tmp/tmp.VkffeZHccK ++ rm /tmp/tmp.ZPwu62rD7o /tmp/tmp.VkffeZHccK ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ytyTuPG3AF +++ mktemp ++ local LAST_ERR=/tmp/tmp.isKwKCni6Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ytyTuPG3AF ++ cat /tmp/tmp.isKwKCni6Z ++ rm /tmp/tmp.ytyTuPG3AF /tmp/tmp.isKwKCni6Z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BFAho2lY4I +++ mktemp ++ local LAST_ERR=/tmp/tmp.ER1YNTdRrq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BFAho2lY4I ++ cat /tmp/tmp.ER1YNTdRrq ++ rm /tmp/tmp.BFAho2lY4I /tmp/tmp.ER1YNTdRrq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MtUPTjaDgU +++ mktemp ++ local LAST_ERR=/tmp/tmp.5fMP6pfSyJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MtUPTjaDgU ++ cat /tmp/tmp.5fMP6pfSyJ ++ rm /tmp/tmp.MtUPTjaDgU /tmp/tmp.5fMP6pfSyJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jkuIpebB8u +++ mktemp ++ local LAST_ERR=/tmp/tmp.9CjbeY6wOf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jkuIpebB8u ++ cat /tmp/tmp.9CjbeY6wOf ++ rm /tmp/tmp.jkuIpebB8u /tmp/tmp.9CjbeY6wOf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gVNRtrJ7RX +++ mktemp ++ local LAST_ERR=/tmp/tmp.7WaaXzOUIg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gVNRtrJ7RX ++ cat /tmp/tmp.7WaaXzOUIg ++ rm /tmp/tmp.gVNRtrJ7RX /tmp/tmp.7WaaXzOUIg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4fT72C2TYm +++ mktemp ++ local LAST_ERR=/tmp/tmp.JPll34NfuR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4fT72C2TYm ++ cat /tmp/tmp.JPll34NfuR ++ rm /tmp/tmp.4fT72C2TYm /tmp/tmp.JPll34NfuR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JE3yOm5cl9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.GlBkJxkvWE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JE3yOm5cl9 ++ cat /tmp/tmp.GlBkJxkvWE ++ rm /tmp/tmp.JE3yOm5cl9 /tmp/tmp.GlBkJxkvWE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mFK3YJmGl1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VC6j9fFlYm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mFK3YJmGl1 ++ cat /tmp/tmp.VC6j9fFlYm ++ rm /tmp/tmp.mFK3YJmGl1 /tmp/tmp.VC6j9fFlYm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H3QET0nvxD +++ mktemp ++ local LAST_ERR=/tmp/tmp.DhaJzDK9p0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.H3QET0nvxD ++ cat /tmp/tmp.DhaJzDK9p0 ++ rm /tmp/tmp.H3QET0nvxD /tmp/tmp.DhaJzDK9p0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IIqFSE5gQ0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.YLhcXrZNmy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IIqFSE5gQ0 ++ cat /tmp/tmp.YLhcXrZNmy ++ rm /tmp/tmp.IIqFSE5gQ0 /tmp/tmp.YLhcXrZNmy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Si2K0hSBOv +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jt6Wjk3IXv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Si2K0hSBOv ++ cat /tmp/tmp.Jt6Wjk3IXv ++ rm /tmp/tmp.Si2K0hSBOv /tmp/tmp.Jt6Wjk3IXv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h9fOZopuee +++ mktemp ++ local LAST_ERR=/tmp/tmp.GHQXIoLjIS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h9fOZopuee ++ cat /tmp/tmp.GHQXIoLjIS ++ rm /tmp/tmp.h9fOZopuee /tmp/tmp.GHQXIoLjIS ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.86XDWM4dKn +++ mktemp ++ local LAST_ERR=/tmp/tmp.UU76dkX8Vq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.86XDWM4dKn ++ cat /tmp/tmp.UU76dkX8Vq ++ rm /tmp/tmp.86XDWM4dKn /tmp/tmp.UU76dkX8Vq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iewfZvl1hV +++ mktemp ++ local LAST_ERR=/tmp/tmp.x1pex27pNg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iewfZvl1hV ++ cat /tmp/tmp.x1pex27pNg ++ rm /tmp/tmp.iewfZvl1hV /tmp/tmp.x1pex27pNg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DHEgD6SGfn +++ mktemp ++ local LAST_ERR=/tmp/tmp.RzhJzRWogR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DHEgD6SGfn ++ cat /tmp/tmp.RzhJzRWogR ++ rm /tmp/tmp.DHEgD6SGfn /tmp/tmp.RzhJzRWogR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Aa0BK4nrzI +++ mktemp ++ local LAST_ERR=/tmp/tmp.h185Eodg7K ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Aa0BK4nrzI ++ cat /tmp/tmp.h185Eodg7K ++ rm /tmp/tmp.Aa0BK4nrzI /tmp/tmp.h185Eodg7K ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hryxKj1d58 +++ mktemp ++ local LAST_ERR=/tmp/tmp.y1zUrg8NF9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hryxKj1d58 ++ cat /tmp/tmp.y1zUrg8NF9 ++ rm /tmp/tmp.hryxKj1d58 /tmp/tmp.y1zUrg8NF9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AAcj6NsdP9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1jvvCxhELM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AAcj6NsdP9 ++ cat /tmp/tmp.1jvvCxhELM ++ rm /tmp/tmp.AAcj6NsdP9 /tmp/tmp.1jvvCxhELM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BjCUVVOVYH +++ mktemp ++ local LAST_ERR=/tmp/tmp.OLtcSZ7d1M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BjCUVVOVYH ++ cat /tmp/tmp.OLtcSZ7d1M ++ rm /tmp/tmp.BjCUVVOVYH /tmp/tmp.OLtcSZ7d1M ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8zcRbnzbge +++ mktemp ++ local LAST_ERR=/tmp/tmp.z8AYxLzpy8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8zcRbnzbge ++ cat /tmp/tmp.z8AYxLzpy8 ++ rm /tmp/tmp.8zcRbnzbge /tmp/tmp.z8AYxLzpy8 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Umy0BKHeVd ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.mhtCNTLyOT +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Umy0BKHeVd +++++ cat /tmp/tmp.mhtCNTLyOT +++++ rm /tmp/tmp.Umy0BKHeVd /tmp/tmp.mhtCNTLyOT +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.nCuO0ULWNf ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.VZfRPBL8ne +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.nCuO0ULWNf +++++ cat /tmp/tmp.VZfRPBL8ne +++++ rm /tmp/tmp.nCuO0ULWNf /tmp/tmp.VZfRPBL8ne +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k9DT0qSa4e +++ mktemp ++ local LAST_ERR=/tmp/tmp.z3M88wrxO2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k9DT0qSa4e ++ cat /tmp/tmp.z3M88wrxO2 ++ rm /tmp/tmp.k9DT0qSa4e /tmp/tmp.z3M88wrxO2 ++ return 0 + [[ 2 == \2 ]] + echo + compare_kubectl secret/pitr-mysql-init + local resource=secret/pitr-mysql-init + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml + local new_result=/tmp/tmp.4FyEIG03fJ/secret_pitr-mysql-init.yml + desc 'compare secret/pitr-mysql-init-' + set +o xtrace ----------------------------------------------------------------------------------- compare secret/pitr-mysql-init- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-84.yml ']' + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.32 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129.yml ']' + version_gt 1.27 ++ echo '1.32 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.32 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k124.yml ']' + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k122.yml ']' + version_gt 1.21 ++ echo '1.32 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-aks.yml ']' + kubectl_bin get -o yaml secret/pitr-mysql-init + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("pitr-31942", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.W8nj0zzPkF ++ mktemp + local LAST_ERR=/tmp/tmp.4KutABtIm6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml secret/pitr-mysql-init + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.W8nj0zzPkF + cat /tmp/tmp.4KutABtIm6 + rm /tmp/tmp.W8nj0zzPkF /tmp/tmp.4KutABtIm6 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml /tmp/tmp.4FyEIG03fJ/secret_pitr-mysql-init.yml + log 'compare_kubectl: secret/pitr-mysql-init OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:02:14+0000]' compare_kubectl: secret/pitr-mysql-init OK [2026-04-29T17:02:14+0000] compare_kubectl: secret/pitr-mysql-init OK + mysql_host=pitr-pxc-0.pitr-pxc.pitr-31942 ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-31942 ++ local host=pitr-pxc-0.pitr-pxc.pitr-31942 +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.CkupqhpnP8 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.BqRUMANnhU ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.CkupqhpnP8 ++++ cat /tmp/tmp.BqRUMANnhU ++++ rm /tmp/tmp.CkupqhpnP8 /tmp/tmp.BqRUMANnhU ++++ return 0 +++ client_pod=pxc-client-56fd5498cd-6jk4n +++ wait_pod pxc-client-56fd5498cd-6jk4n +++ local pod=pxc-client-56fd5498cd-6jk4n +++ local max_retry=480 +++ local ns= ++++ grep -E '^(pxc|proxysql)$' ++++ echo pxc-client-56fd5498cd-6jk4n ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' +++ local container= +++ set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local gtid=e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26 ++ [[ ! e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26 + gtid=e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26 + log '[CASE 1]: gtid_executed: e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:02:28+0000]' '[CASE' '1]:' gtid_executed: e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26 [2026-04-29T17:02:28+0000] [CASE 1]: gtid_executed: e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26 + write_data_for_pitr pitr 100503 3 + local cluster=pitr + local start=100503 + local rows=3 ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pVy0pBgdkZ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.7sCvJsZDy3 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.pVy0pBgdkZ +++ cat /tmp/tmp.7sCvJsZDy3 +++ rm /tmp/tmp.pVy0pBgdkZ /tmp/tmp.7sCvJsZDy3 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6kZkgVYTkk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.G71k2V3gjC +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.6kZkgVYTkk +++ cat /tmp/tmp.G71k2V3gjC +++ rm /tmp/tmp.6kZkgVYTkk /tmp/tmp.G71k2V3gjC +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + local sql= + (( i=0 )) + (( i/e3794f29-43ea-11f1-a5b1-32c5d2a852bf:26/g' + /usr/bin/sed -e 's///g' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/restore-on-pitr-minio-gtid.yaml + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-31942~ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.AyJF7mQby1 ++ mktemp + local LAST_ERR=/tmp/tmp.W1dhpmdBH2 + local exit_status=0 + /usr/bin/sed -e 's///g' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.AyJF7mQby1 perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-gtid created + cat /tmp/tmp.W1dhpmdBH2 + rm /tmp/tmp.AyJF7mQby1 /tmp/tmp.W1dhpmdBH2 + return 0 + wait_backup_restore restore-on-pitr-minio-gtid 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Stopping Cluster state 2026-04-29T17:04:04 pxc-restore/restore-on-pitr-minio-gtid state: Stopping Cluster + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ......Error from server (NotFound): pods "pitr-pxc-1" not found + wait_for_delete pod/pitr-pxc-0 + local res=pod/pitr-pxc-0 + echo -n 'waiting for pod/pitr-pxc-0 to be deleted' waiting for pod/pitr-pxc-0 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-0" not found + wait_backup_restore restore-on-pitr-minio-gtid 'Point-in-time recovering' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Point-in-time recovering' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Point-in-time recovering state 2026-04-29T17:04:47 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:04:50 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:04:54 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:04:57 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:04:59 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:03 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:06 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:11 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:13 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:17 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:20 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:23 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:27 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:29 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:33 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:35 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:40 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:44 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:47 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:50 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:53 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:56 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:05:58 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:06:01 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:06:03 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:06:06 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:06:09 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:06:12 pxc-restore/restore-on-pitr-minio-gtid state: Point-in-time recovering + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-1" not found + wait_pod pitr-pxc-0 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-pxc-0 + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + wait_backup_restore restore-on-pitr-minio-gtid 'Starting Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Starting Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Starting Cluster state 2026-04-29T17:06:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zFlST04vj1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yeggs1wNAg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zFlST04vj1 ++ cat /tmp/tmp.yeggs1wNAg ++ rm /tmp/tmp.zFlST04vj1 /tmp/tmp.yeggs1wNAg ++ return 0 + local minio_pod=minio-service-757b74c96c-8xdn6 + kubectl_bin exec minio-service-757b74c96c-8xdn6 -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json ++ mktemp + local LAST_OUT=/tmp/tmp.jNX0AErDoU ++ mktemp + local LAST_ERR=/tmp/tmp.vidkvtWG8F + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec minio-service-757b74c96c-8xdn6 -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl exec minio-service-757b74c96c-8xdn6 -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl exec minio-service-757b74c96c-8xdn6 -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.jNX0AErDoU + cat /tmp/tmp.vidkvtWG8F stat: cannot statx '/exports/operator-testing/binlogs/gtid-binlog-cache.json': No such file or directory command terminated with exit code 1 + rm /tmp/tmp.jNX0AErDoU /tmp/tmp.vidkvtWG8F + return 1 + wait_backup_restore restore-on-pitr-minio-gtid + local backup_name=restore-on-pitr-minio-gtid + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Succeeded state 2026-04-29T17:06:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:06:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:06:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:06:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:06:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:06:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:06:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:06:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:07:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:08:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:09:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:09:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:09:08 pxc-restore/restore-on-pitr-minio-gtid state: Succeeded + kubectl_bin logs job/restore-job-restore-on-pitr-minio-gtid-pitr ++ mktemp + local LAST_OUT=/tmp/tmp.BKrAJL9RyN ++ mktemp + local LAST_ERR=/tmp/tmp.b9MIoA367V + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs job/restore-job-restore-on-pitr-minio-gtid-pitr + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BKrAJL9RyN + LIB_PATH=/opt/percona/backup/lib/pxc + . /opt/percona/backup/lib/pxc/check-version.sh + . /opt/percona/backup/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + . /opt/percona/backup/lib/pxc/aws.sh ++ set -o errexit ++ export AWS_SHARED_CREDENTIALS_FILE=/tmp/aws-credfile ++ AWS_SHARED_CREDENTIALS_FILE=/tmp/aws-credfile ++ export AWS_REGION=us-east-1 ++ AWS_REGION=us-east-1 ++ export AWS_ENDPOINT_URL=https://minio-service.pitr-31942:9000/ ++ AWS_ENDPOINT_URL=https://minio-service.pitr-31942:9000/ ++ '[' -n false ']' ++ [[ false == \f\a\l\s\e ]] ++ AWS_S3_NO_VERIFY_SSL=--no-verify-ssl ++ caBundleDir=/etc/s3/certs ++ caBundleFile=/etc/s3/certs/ca.crt ++ '[' -f /etc/s3/certs/ca.crt ']' + XBCLOUD_ARGS='--curl-retriable-errors=7 ' + '[' -n false ']' + [[ false == \f\a\l\s\e ]] + XBCLOUD_ARGS='--insecure --curl-retriable-errors=7 ' + '[' -n operator-testing/pitr-2026-04-29-16:51:07-full ']' + set -x + aws --no-verify-ssl s3 ls operator-testing/pitr-2026-04-29-16:51:07-full urllib3/connectionpool.py:1097: InsecureRequestWarning: Unverified HTTPS request is being made to host 'minio-service.pitr-31942'. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/latest/advanced-usage.html#tls-warnings PRE pitr-2026-04-29-16:51:07-full.sst_info/ PRE pitr-2026-04-29-16:51:07-full/ 2026-04-29 16:51:36 27362 pitr-2026-04-29-16:51:07-full.md5 2026-04-29 16:51:21 128 pitr-2026-04-29-16:51:07-full.sst_info.md5 + '[' -n '' ']' + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000001 /datadir/binlog.000002 /datadir/binlog.000003 /datadir/binlog.000004 /datadir/binlog.000005 /datadir/binlog.000006 /datadir/binlog.000007 /datadir/binlog.000008 /datadir/binlog.000009 /datadir/binlog.000010 /datadir/binlog.000011 /datadir/binlog.000012 /datadir/binlog.000013 /datadir/binlog.000014 /datadir/binlog.000015 /datadir/binlog.000016 /datadir/binlog.000017 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql-state-monitor /datadir/mysql-state-monitor.log /datadir/mysql.ibd /datadir/mysql.state /datadir/mysql_upgrade_history /datadir/notify.sock /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/prepare_restored_cluster.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sys /datadir/test /datadir/undo_001 /datadir/undo_002 /datadir/version_info /datadir/wsrep_cmd_notify_handler.sh /datadir/xtrabackup-server-sidecar ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_KSGe ++ grep -c processor /proc/cpuinfo ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_KSGe --parallel=4 ++ destination ++ '[' -n operator-testing/pitr-2026-04-29-16:51:07-full ']' ++ echo -n s3://operator-testing/pitr-2026-04-29-16:51:07-full + xbcloud get --parallel=4 --insecure --curl-retriable-errors=7 s3://operator-testing/pitr-2026-04-29-16:51:07-full.sst_info error: http request failed: Couldn't resolve host name error: http request failed: Couldn't resolve host name 260429 17:05:14 xbcloud: Successfully connected. 260429 17:05:14 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full.sst_info/sst_info.00000000000000000000. 260429 17:05:14 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full.sst_info/sst_info.00000000000000000000, size 141 260429 17:05:14 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full.sst_info/sst_info.00000000000000000001. 260429 17:05:14 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full.sst_info/sst_info.00000000000000000001, size 22 260429 17:05:14 xbcloud: Download completed. ++ get_xtrabackup_version +++ xtrabackup --version +++ grep '^xtrabackup version' +++ sed 's/-.*//' +++ awk '{print $3}' ++ xtrabackup_ver=8.4.0 ++ echo 8.4.0 + XTRABACKUP_VERSION=8.4.0 + check_for_version 8.4.0 8.0.0 + '[' -z 8.4.0 ']' + '[' -z 8.0.0 ']' + local local_version_str + local required_version_str ++ normalize_version 8.4.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.4.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=4 ++ patch=0 ++ printf %02d%02d%02d 8 4 0 + local_version_str=080400 ++ normalize_version 8.0.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=0 ++ printf %02d%02d%02d 8 0 0 + required_version_str=080000 + [[ 080400 < 080000 ]] + return 0 + XBSTREAM_EXTRA_ARGS=' --decompress' ++ grep -c processor /proc/cpuinfo ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_KSGe --parallel=4 --decompress ++ destination ++ '[' -n operator-testing/pitr-2026-04-29-16:51:07-full ']' ++ echo -n s3://operator-testing/pitr-2026-04-29-16:51:07-full + xbcloud get --parallel=4 --insecure --curl-retriable-errors=7 s3://operator-testing/pitr-2026-04-29-16:51:07-full error: http request failed: Couldn't resolve host name error: http request failed: Couldn't resolve host name 260429 17:05:15 xbcloud: Successfully connected. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_tablespaces.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_logfile.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_info.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_checkpoints.00000000000000000000. 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_tablespaces.lz4.00000000000000000000, size 128 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_logfile.lz4.00000000000000000000, size 360 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_checkpoints.00000000000000000000, size 183 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_info.lz4.00000000000000000000, size 862 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_checkpoints.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_info.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_logfile.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_tablespaces.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_checkpoints.00000000000000000001, size 36 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_info.lz4.00000000000000000001, size 33 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_tablespaces.lz4.00000000000000000001, size 40 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_logfile.lz4.00000000000000000001, size 36 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/variables_info_196.sdi.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/users_154.sdi.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/user_variables_b_186.sdi.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_cur_92.sdi.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_cur_92.sdi.lz4.00000000000000000000, size 2271 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/users_154.sdi.lz4.00000000000000000000, size 1497 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/variables_info_196.sdi.lz4.00000000000000000000, size 1880 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/user_variables_b_186.sdi.lz4.00000000000000000000, size 1593 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_cur_92.sdi.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/user_variables_b_186.sdi.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/users_154.sdi.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/variables_info_196.sdi.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_cur_92.sdi.lz4.00000000000000000001, size 60 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/user_variables_b_186.sdi.lz4.00000000000000000001, size 61 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/users_154.sdi.lz4.00000000000000000001, size 50 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/variables_info_196.sdi.lz4.00000000000000000001, size 59 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/sys/sys_config.ibd.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/rwlock_instances_109.sdi.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000000. 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000000, size 1767 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/sys/sys_config.ibd.lz4.00000000000000000000, size 2722 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/rwlock_instances_109.sdi.lz4.00000000000000000000, size 1643 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000000, size 1769 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/rwlock_instances_109.sdi.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/sys/sys_config.ibd.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000001. 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/rwlock_instances_109.sdi.lz4.00000000000000000001, size 61 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/sys/sys_config.ibd.lz4.00000000000000000001, size 36 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000001, size 61 260429 17:05:15 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/variables_by_thr_193.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/binary_log_trans_199.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/variables_by_thr_193.sdi.lz4.00000000000000000000, size 1566 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000000, size 1716 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/binary_log_trans_199.sdi.lz4.00000000000000000000, size 2080 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000000, size 2567 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/binary_log_trans_199.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/variables_by_thr_193.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/binary_log_trans_199.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/variables_by_thr_193.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000000, size 2168 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000000, size 2039 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000000, size 3068 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000000, size 1833 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_128.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_128.sdi.lz4.00000000000000000000, size 1518 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000000, size 1672 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000000, size 1612 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000000, size 1957 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_128.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_128.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_123.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_consumers_111.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/keyring_keys_162.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000000, size 1909 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/keyring_keys_162.sdi.lz4.00000000000000000000, size 1277 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_consumers_111.sdi.lz4.00000000000000000000, size 1440 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_123.sdi.lz4.00000000000000000000, size 1767 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/keyring_keys_162.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_consumers_111.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_123.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/keyring_keys_162.sdi.lz4.00000000000000000001, size 57 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_hi_123.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_consumers_111.sdi.lz4.00000000000000000001, size 60 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/slow_log_228.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/cond_instances_90.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_159.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/cond_instances_90.sdi.lz4.00000000000000000000, size 1461 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/slow_log_228.sdi.lz4.00000000000000000000, size 1714 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000000, size 1607 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_159.sdi.lz4.00000000000000000000, size 1779 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_159.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/cond_instances_90.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/slow_log_228.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/cond_instances_90.sdi.lz4.00000000000000000001, size 58 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_159.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/slow_log_228.sdi.lz4.00000000000000000001, size 40 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/general_log_227.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000000, size 2738 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000000, size 1767 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000000, size 1554 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/general_log_227.sdi.lz4.00000000000000000000, size 1491 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/general_log_227.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/general_log_227.sdi.lz4.00000000000000000001, size 43 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_95.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/backup-my.cnf.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/metadata_locks_169.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_95.sdi.lz4.00000000000000000000, size 1609 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000000, size 2487 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/metadata_locks_169.sdi.lz4.00000000000000000000, size 1851 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/backup-my.cnf.lz4.00000000000000000000, size 430 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/metadata_locks_169.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/backup-my.cnf.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_95.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/metadata_locks_169.sdi.lz4.00000000000000000001, size 59 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000001, size 47 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/backup-my.cnf.lz4.00000000000000000001, size 31 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_95.sdi.lz4.00000000000000000001, size 60 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_host_188.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_cu_121.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/keyring_componen_203.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/keyring_componen_203.sdi.lz4.00000000000000000000, size 1269 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_cu_121.sdi.lz4.00000000000000000000, size 1909 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_host_188.sdi.lz4.00000000000000000000, size 1468 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000000, size 2204 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/keyring_componen_203.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_cu_121.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_host_188.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000001, size 60 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_stages_cu_121.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/keyring_componen_203.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_host_188.sdi.lz4.00000000000000000001, size 59 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSV.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_binlog_info.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_148.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSV.lz4.00000000000000000000, size 79 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_binlog_info.lz4.00000000000000000000, size 148 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000000, size 2540 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_148.sdi.lz4.00000000000000000000, size 1671 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/xtrabackup_binlog_info.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_148.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSV.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_148.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSV.lz4.00000000000000000001, size 36 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/xtrabackup_binlog_info.lz4.00000000000000000001, size 40 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000001, size 41 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/general_log.CSM.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_202.sdi.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSM.lz4.00000000000000000000. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_202.sdi.lz4.00000000000000000000, size 1276 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSM.lz4.00000000000000000000, size 96 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000000, size 2102 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/general_log.CSM.lz4.00000000000000000000, size 99 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSM.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_202.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/general_log.CSM.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000001. 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/slow_log.CSM.lz4.00000000000000000001, size 36 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_202.sdi.lz4.00000000000000000001, size 57 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000001, size 61 260429 17:05:16 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/general_log.CSM.lz4.00000000000000000001, size 39 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql.ibd.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/myApp/myApp.ibd.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/myApp/myApp.ibd.lz4.00000000000000000000, size 2322 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000000, size 1731 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000000, size 2796 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql.ibd.lz4.00000000000000000000, size 4521234 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/myApp/myApp.ibd.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql.ibd.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/myApp/myApp.ibd.lz4.00000000000000000001, size 33 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql.ibd.lz4.00000000000000000001, size 27 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000001, size 49 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/account_failed_l_204.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_179.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/ib_buffer_pool.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_179.sdi.lz4.00000000000000000000, size 1708 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/ib_buffer_pool.lz4.00000000000000000000, size 1234 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/account_failed_l_204.sdi.lz4.00000000000000000000, size 1815 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000000, size 1695 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/ib_buffer_pool.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_179.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/account_failed_l_204.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/account_failed_l_204.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_179.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/ib_buffer_pool.lz4.00000000000000000001, size 32 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/global_variables_194.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/binlog.index.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/undo_002.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000000, size 1555 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/binlog.index.lz4.00000000000000000000, size 95 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/global_variables_194.sdi.lz4.00000000000000000000, size 1450 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/undo_002.lz4.00000000000000000000, size 1899969 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/undo_002.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/binlog.index.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/global_variables_194.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/undo_002.lz4.00000000000000000001, size 26 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/binlog.index.lz4.00000000000000000001, size 30 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/global_variables_194.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/general_log.CSV.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/ibdata1.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/file_instances_101.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000000, size 1705 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/file_instances_101.sdi.lz4.00000000000000000000, size 1494 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/general_log.CSV.lz4.00000000000000000000, size 82 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/ibdata1.lz4.00000000000000000000, size 56347 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/file_instances_101.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/ibdata1.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/mysql/general_log.CSV.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/file_instances_101.sdi.lz4.00000000000000000001, size 59 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/ibdata1.lz4.00000000000000000001, size 25 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/mysql/general_log.CSV.lz4.00000000000000000001, size 39 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_status_192.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_threads_116.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/accounts_155.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/accounts_155.sdi.lz4.00000000000000000000, size 1561 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000000, size 1677 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_threads_116.sdi.lz4.00000000000000000000, size 1742 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_status_192.sdi.lz4.00000000000000000000, size 1440 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/accounts_155.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_threads_116.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_status_192.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_status_192.sdi.lz4.00000000000000000001, size 59 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_threads_116.sdi.lz4.00000000000000000001, size 58 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/accounts_155.sdi.lz4.00000000000000000001, size 53 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_account__161.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000000, size 1996 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000000, size 3072 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_account__161.sdi.lz4.00000000000000000000, size 1636 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000000, size 2105 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_account__161.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_account__161.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/data_lock_waits_171.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_93.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/error_log_91.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000000, size 1796 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/data_lock_waits_171.sdi.lz4.00000000000000000000, size 1926 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/error_log_91.sdi.lz4.00000000000000000000, size 1990 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_93.sdi.lz4.00000000000000000000, size 2271 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/error_log_91.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_93.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/data_lock_waits_171.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/error_log_91.sdi.lz4.00000000000000000001, size 53 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_his_93.sdi.lz4.00000000000000000001, size 60 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/data_lock_waits_171.sdi.lz4.00000000000000000001, size 60 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_meters_113.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_153.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000000, size 1740 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_153.sdi.lz4.00000000000000000000, size 1649 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000000, size 2273 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_meters_113.sdi.lz4.00000000000000000000, size 1585 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_153.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_meters_113.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_153.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_meters_113.sdi.lz4.00000000000000000001, size 57 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/mutex_instances_105.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000000, size 2911 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000000, size 1594 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/mutex_instances_105.sdi.lz4.00000000000000000000, size 1581 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000000, size 1521 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/mutex_instances_105.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/mutex_instances_105.sdi.lz4.00000000000000000001, size 60 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000001, size 60 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/socket_instances_157.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__102.sdi.lz4.00000000000000000000. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000000, size 1556 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/socket_instances_157.sdi.lz4.00000000000000000000, size 1837 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000000, size 1695 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__102.sdi.lz4.00000000000000000000, size 1810 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/socket_instances_157.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__102.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000001. 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000001, size 60 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/socket_instances_157.sdi.lz4.00000000000000000001, size 61 260429 17:05:17 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__102.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/undo_001.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000000, size 1614 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000000, size 1963 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000000, size 2737 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/undo_001.lz4.00000000000000000000, size 152028 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/undo_001.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/undo_001.lz4.00000000000000000001, size 26 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000001, size 60 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/host_cache_104.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/log_status_184.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_actors_110.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000000, size 1674 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_actors_110.sdi.lz4.00000000000000000000, size 1567 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/host_cache_104.sdi.lz4.00000000000000000000, size 2301 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/log_status_184.sdi.lz4.00000000000000000000, size 1293 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_actors_110.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/log_status_184.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/host_cache_104.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/log_status_184.sdi.lz4.00000000000000000001, size 55 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_actors_110.sdi.lz4.00000000000000000001, size 57 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/host_cache_104.sdi.lz4.00000000000000000001, size 55 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000001, size 60 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_tot_201.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_167.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_g_163.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/objects_summary__106.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_tot_201.sdi.lz4.00000000000000000000, size 1256 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_167.sdi.lz4.00000000000000000000, size 1734 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_g_163.sdi.lz4.00000000000000000000, size 1632 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/objects_summary__106.sdi.lz4.00000000000000000000, size 1543 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/objects_summary__106.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_g_163.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_167.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_tot_201.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_b_167.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/objects_summary__106.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/malloc_stats_tot_201.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/memory_summary_g_163.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/performance_time_107.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/persisted_variab_197.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/prepared_stateme_185.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/processlist_108.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/persisted_variab_197.sdi.lz4.00000000000000000000, size 1448 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/prepared_stateme_185.sdi.lz4.00000000000000000000, size 2792 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/performance_time_107.sdi.lz4.00000000000000000000, size 1467 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/processlist_108.sdi.lz4.00000000000000000000, size 1980 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/processlist_108.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/prepared_stateme_185.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/persisted_variab_197.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/performance_time_107.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/prepared_stateme_185.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/processlist_108.sdi.lz4.00000000000000000001, size 56 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/persisted_variab_197.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/performance_time_107.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_handles_168.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/pxc_cluster_view_205.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_accoun_187.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/threads_120.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_accoun_187.sdi.lz4.00000000000000000000, size 1537 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_handles_168.sdi.lz4.00000000000000000000, size 1737 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/pxc_cluster_view_205.sdi.lz4.00000000000000000000, size 1344 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/threads_120.sdi.lz4.00000000000000000000, size 2721 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/threads_120.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_accoun_187.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/pxc_cluster_view_205.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_handles_168.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/threads_120.sdi.lz4.00000000000000000001, size 52 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_accoun_187.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_handles_168.sdi.lz4.00000000000000000001, size 58 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/pxc_cluster_view_205.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_140.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000000, size 2393 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000000, size 2135 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_140.sdi.lz4.00000000000000000000, size 1691 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000000, size 2326 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_140.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/events_statement_140.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_178.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/data_locks_170.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000000, size 1876 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_178.sdi.lz4.00000000000000000000, size 2441 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/data_locks_170.sdi.lz4.00000000000000000000, size 2030 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000000, size 1742 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_178.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/data_locks_170.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/data_locks_170.sdi.lz4.00000000000000000001, size 55 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_178.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_181.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_183.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_172.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_181.sdi.lz4.00000000000000000000, size 1522 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_183.sdi.lz4.00000000000000000000, size 1721 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_172.sdi.lz4.00000000000000000000, size 3057 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000000, size 1995 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_172.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_183.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_181.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_172.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_asyn_183.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_appl_181.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_173.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_connect__160.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_174.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_variable_195.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_173.sdi.lz4.00000000000000000000, size 1511 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_variable_195.sdi.lz4.00000000000000000000, size 1447 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_174.sdi.lz4.00000000000000000000, size 2554 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_connect__160.sdi.lz4.00000000000000000000, size 1631 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_variable_195.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_174.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/session_connect__160.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_173.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_variable_195.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/session_connect__160.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_conn_174.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/replication_grou_173.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/binlog.000009.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_instrument_112.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_metrics_114.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_objects_115.sdi.lz4.00000000000000000000. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_instrument_112.sdi.lz4.00000000000000000000, size 2021 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/binlog.000009.lz4.00000000000000000000, size 224 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_metrics_114.sdi.lz4.00000000000000000000, size 1728 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_objects_115.sdi.lz4.00000000000000000000, size 1719 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_objects_115.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_metrics_114.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/setup_instrument_112.sdi.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/binlog.000009.lz4.00000000000000000001. 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_objects_115.sdi.lz4.00000000000000000001, size 58 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_instrument_112.sdi.lz4.00000000000000000001, size 61 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/setup_metrics_114.sdi.lz4.00000000000000000001, size 58 260429 17:05:18 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/binlog.000009.lz4.00000000000000000001, size 31 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_thread_189.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/user_defined_fun_198.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_user_190.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/global_status_191.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_thread_189.sdi.lz4.00000000000000000000, size 1565 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_user_190.sdi.lz4.00000000000000000000, size 1474 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/user_defined_fun_198.sdi.lz4.00000000000000000000, size 1548 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/global_status_191.sdi.lz4.00000000000000000000, size 1443 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/global_status_191.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_user_190.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/user_defined_fun_198.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/status_by_thread_189.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/global_status_191.sdi.lz4.00000000000000000001, size 58 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_user_190.sdi.lz4.00000000000000000001, size 59 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/status_by_thread_189.sdi.lz4.00000000000000000001, size 61 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/user_defined_fun_198.sdi.lz4.00000000000000000001, size 61 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/hosts_156.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_118.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_lock_waits_119.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/tls_channel_stat_200.sdi.lz4.00000000000000000000. 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/hosts_156.sdi.lz4.00000000000000000000, size 1493 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_118.sdi.lz4.00000000000000000000, size 1929 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/tls_channel_stat_200.sdi.lz4.00000000000000000000, size 1301 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_lock_waits_119.sdi.lz4.00000000000000000000, size 2308 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/tls_channel_stat_200.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_lock_waits_119.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_118.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Downloading pitr-2026-04-29-16:51:07-full/performance_schema/hosts_156.sdi.lz4.00000000000000000001. 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/hosts_156.sdi.lz4.00000000000000000001, size 50 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_lock_waits_119.sdi.lz4.00000000000000000001, size 61 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/table_io_waits_s_118.sdi.lz4.00000000000000000001, size 61 260429 17:05:19 xbcloud: [0] Download successfull pitr-2026-04-29-16:51:07-full/performance_schema/tls_channel_stat_200.sdi.lz4.00000000000000000001, size 61 260429 17:05:19 xbcloud: Download completed. + set +o xtrace vault configuration not found + xtrabackup --use-memory=100MB --prepare --rollback-prepared-trx --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_KSGe 2026-04-29T17:05:19.261408-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=18958992 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2026-04-29T17:05:19.261484-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=100MB --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_KSGe xtrabackup version 8.4.0-5 based on MySQL server 8.4.0 Linux (x86_64) (revision id: c6054e17) 2026-04-29T17:05:19.261522-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_KSGe/ 2026-04-29T17:05:19.261607-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2026-04-29T17:05:19.271135-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(30295711) 2026-04-29T17:05:19.272306-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-04-29T17:05:19.272329-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-04-29T17:05:19.272336-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-04-29T17:05:19.272360-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-04-29T17:05:19.272370-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-04-29T17:05:19.272378-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-04-29T17:05:19.272568-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2026-04-29T17:05:19.475819-00:00 0 [Note] [MY-011825] [Xtrabackup] Generating a list of tablespaces 2026-04-29T17:05:19.475853-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2026-04-29T17:05:19.478773-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 2 files. 2026-04-29T17:05:19.483593-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2026-04-29T17:05:19.492609-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2026-04-29T17:05:19.493606-00:00 0 [Note] [MY-013795] [InnoDB] Encryption key is loaded for undo tablespace 'innodb_undo_001'. 2026-04-29T17:05:19.493680-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2026-04-29T17:05:19.494729-00:00 0 [Note] [MY-013795] [InnoDB] Encryption key is loaded for undo tablespace 'innodb_undo_002'. 2026-04-29T17:05:19.495313-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2026-04-29T17:05:19.696830-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-04-29T17:05:19.696842-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-04-29T17:05:19.696846-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-04-29T17:05:19.696854-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-04-29T17:05:19.696860-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-04-29T17:05:19.696866-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-04-29T17:05:19.696943-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-04-29T17:05:19.696955-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-04-29T17:05:19.696961-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-04-29T17:05:19.696969-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-04-29T17:05:19.696975-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-04-29T17:05:19.696979-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-04-29T17:05:19.696990-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2026-04-29T17:05:19.696999-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 104857600 bytes for buffer pool (set by --use-memory parameter) 2026-04-29T17:05:19.697013-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2026-04-29T17:05:19.697019-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2026-04-29T17:05:19.697023-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2026-04-29T17:05:19.697027-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2026-04-29T17:05:19.697122-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2026-04-29T17:05:19.697424-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2026-04-29T17:05:19.697460-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2026-04-29T17:05:19.698473-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2026-04-29T17:05:19.698698-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 128.000000M, instances = 1, chunk size =128.000000M 2026-04-29T17:05:19.705245-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2026-04-29T17:05:19.706822-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2026-04-29T17:05:19.733854-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 30295711 in redo log file ./#innodb_redo/#ib_redo0. 2026-04-29T17:05:19.733900-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 30144388 in the system tablespace does not match the log sequence number 30295711 in the redo log files! 2026-04-29T17:05:19.733914-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2026-04-29T17:05:19.733921-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2026-04-29T17:05:19.734038-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 30295598, whereas checkpoint_lsn = 30295711 and start_lsn = 30295552 2026-04-29T17:05:19.734050-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 30295731 2026-04-29T17:05:19.751855-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-04-29T17:05:19.752495-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2026-04-29T17:05:19.752546-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2026-04-29T17:05:19.752559-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2026-04-29T17:05:19.855926-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-04-29T17:05:19.857463-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=30295731. 2026-04-29T17:05:19.857473-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2026-04-29T17:05:19.857507-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2026-04-29T17:05:19.886897-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2026-04-29T17:05:19.892782-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2026-04-29T17:05:19.898056-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=30296076 2026-04-29T17:05:19.898128-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-04-29T17:05:19.899338-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2026-04-29T17:05:19.900411-00:00 0 [Note] [MY-013795] [InnoDB] Encryption key is loaded for undo tablespace 'innodb_undo_001'. 2026-04-29T17:05:19.900482-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2026-04-29T17:05:19.901445-00:00 0 [Note] [MY-013795] [InnoDB] Encryption key is loaded for undo tablespace 'innodb_undo_002'. 2026-04-29T17:05:19.901815-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2026-04-29T17:05:19.901864-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5261 2026-04-29T17:05:20.043355-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2026-04-29T17:05:20.043374-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 141 ms. 2026-04-29T17:05:20.043427-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2026-04-29T17:05:20.043488-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2026-04-29T17:05:20.075314-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2026-04-29T17:05:20.075442-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2026-04-29T17:05:20.109638-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2026-04-29T17:05:20.109792-00:00 0 [Note] [MY-012976] [InnoDB] 8.4.0 started; log sequence number 30296086 2026-04-29T17:05:20.110808-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2026-04-29T17:05:20.116113-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00629566 seconds 2026-04-29T17:05:20.138072-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.0219421 seconds 2026-04-29T17:05:21.138499-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: e3794f29-43ea-11f1-a5b1-32c5d2a852bf:42 2026-04-29T17:05:21.138541-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2026-04-29T17:05:21.138592-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2026-04-29T17:05:22.138454-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2026-04-29T17:05:22.239296-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-04-29T17:05:22.251942-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 30296086 2026-04-29T17:05:22.254623-00:00 0 [Note] [MY-015019] [Server] MySQL Server: Plugins Shutdown - start. 2026-04-29T17:05:22.254635-00:00 0 [Note] [MY-015020] [Server] MySQL Server: Plugins Shutdown - end. 2026-04-29T17:05:22.254897-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --force-non-empty-directories --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_KSGe 2026-04-29T17:05:22.277562-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2026-04-29T17:05:22.277629-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_KSGe xtrabackup version 8.4.0-5 based on MySQL server 8.4.0 Linux (x86_64) (revision id: c6054e17) 2026-04-29T17:05:22.277663-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_KSGe/ 2026-04-29T17:05:22.278230-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2026-04-29T17:05:22.278264-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2026-04-29T17:05:22.278295-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2026-04-29T17:05:22.278345-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2026-04-29T17:05:22.278558-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2026-04-29T17:05:22.278590-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2026-04-29T17:05:22.278848-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000009 to /datadir//binlog.000009 2026-04-29T17:05:22.278877-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000009 to /datadir//binlog.000009 2026-04-29T17:05:22.278959-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2026-04-29T17:05:22.278989-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2026-04-29T17:05:22.279377-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-04-29T17:05:22.279421-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-04-29T17:05:22.279447-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-04-29T17:05:22.279468-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-04-29T17:05:22.279489-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_228.sdi to /datadir/mysql/slow_log_228.sdi 2026-04-29T17:05:22.279508-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_228.sdi to /datadir/mysql/slow_log_228.sdi 2026-04-29T17:05:22.279530-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-04-29T17:05:22.279549-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-04-29T17:05:22.279570-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-04-29T17:05:22.279671-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-04-29T17:05:22.279704-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-04-29T17:05:22.279725-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-04-29T17:05:22.279767-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_227.sdi to /datadir/mysql/general_log_227.sdi 2026-04-29T17:05:22.279787-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_227.sdi to /datadir/mysql/general_log_227.sdi 2026-04-29T17:05:22.279809-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-04-29T17:05:22.279832-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-04-29T17:05:22.279853-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-04-29T17:05:22.279872-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-04-29T17:05:22.279911-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2026-04-29T17:05:22.279933-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2026-04-29T17:05:22.279956-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-04-29T17:05:22.279983-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-04-29T17:05:22.279992-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2026-04-29T17:05:22.280025-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2026-04-29T17:05:22.280049-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-04-29T17:05:22.280066-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-04-29T17:05:22.280100-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2026-04-29T17:05:22.280119-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2026-04-29T17:05:22.280135-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2026-04-29T17:05:22.280154-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2026-04-29T17:05:22.280233-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-04-29T17:05:22.280264-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-04-29T17:05:22.280296-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-04-29T17:05:22.280339-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-04-29T17:05:22.280369-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-04-29T17:05:22.280394-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-04-29T17:05:22.280423-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_168.sdi to /datadir/performance_schema/table_handles_168.sdi 2026-04-29T17:05:22.280449-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_168.sdi to /datadir/performance_schema/table_handles_168.sdi 2026-04-29T17:05:22.280481-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-04-29T17:05:22.280505-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-04-29T17:05:22.280535-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_187.sdi to /datadir/performance_schema/status_by_accoun_187.sdi 2026-04-29T17:05:22.280563-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_187.sdi to /datadir/performance_schema/status_by_accoun_187.sdi 2026-04-29T17:05:22.280592-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_178.sdi to /datadir/performance_schema/replication_appl_178.sdi 2026-04-29T17:05:22.280618-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_178.sdi to /datadir/performance_schema/replication_appl_178.sdi 2026-04-29T17:05:22.280648-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-04-29T17:05:22.280677-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-04-29T17:05:22.280706-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_201.sdi to /datadir/performance_schema/malloc_stats_tot_201.sdi 2026-04-29T17:05:22.280734-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_201.sdi to /datadir/performance_schema/malloc_stats_tot_201.sdi 2026-04-29T17:05:22.280757-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-04-29T17:05:22.280779-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-04-29T17:05:22.280808-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-04-29T17:05:22.280837-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-04-29T17:05:22.280868-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-04-29T17:05:22.280893-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-04-29T17:05:22.280921-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-04-29T17:05:22.280946-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-04-29T17:05:22.280973-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-04-29T17:05:22.280991-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-04-29T17:05:22.281008-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_156.sdi to /datadir/performance_schema/hosts_156.sdi 2026-04-29T17:05:22.281027-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_156.sdi to /datadir/performance_schema/hosts_156.sdi 2026-04-29T17:05:22.281052-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_198.sdi to /datadir/performance_schema/user_defined_fun_198.sdi 2026-04-29T17:05:22.281080-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_198.sdi to /datadir/performance_schema/user_defined_fun_198.sdi 2026-04-29T17:05:22.281116-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-04-29T17:05:22.281145-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-04-29T17:05:22.281177-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_148.sdi to /datadir/performance_schema/events_transacti_148.sdi 2026-04-29T17:05:22.281198-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_148.sdi to /datadir/performance_schema/events_transacti_148.sdi 2026-04-29T17:05:22.281228-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_meters_113.sdi to /datadir/performance_schema/setup_meters_113.sdi 2026-04-29T17:05:22.281256-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_meters_113.sdi to /datadir/performance_schema/setup_meters_113.sdi 2026-04-29T17:05:22.281287-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_163.sdi to /datadir/performance_schema/memory_summary_g_163.sdi 2026-04-29T17:05:22.281330-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_163.sdi to /datadir/performance_schema/memory_summary_g_163.sdi 2026-04-29T17:05:22.281361-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_107.sdi to /datadir/performance_schema/performance_time_107.sdi 2026-04-29T17:05:22.281390-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_107.sdi to /datadir/performance_schema/performance_time_107.sdi 2026-04-29T17:05:22.281420-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_173.sdi to /datadir/performance_schema/replication_grou_173.sdi 2026-04-29T17:05:22.281448-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_173.sdi to /datadir/performance_schema/replication_grou_173.sdi 2026-04-29T17:05:22.281477-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_179.sdi to /datadir/performance_schema/replication_grou_179.sdi 2026-04-29T17:05:22.281502-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_179.sdi to /datadir/performance_schema/replication_grou_179.sdi 2026-04-29T17:05:22.281534-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_153.sdi to /datadir/performance_schema/events_errors_su_153.sdi 2026-04-29T17:05:22.281561-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_153.sdi to /datadir/performance_schema/events_errors_su_153.sdi 2026-04-29T17:05:22.281588-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-04-29T17:05:22.281615-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-04-29T17:05:22.281646-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-04-29T17:05:22.281671-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-04-29T17:05:22.281702-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_101.sdi to /datadir/performance_schema/file_instances_101.sdi 2026-04-29T17:05:22.281729-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_101.sdi to /datadir/performance_schema/file_instances_101.sdi 2026-04-29T17:05:22.281760-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-04-29T17:05:22.281788-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-04-29T17:05:22.281816-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_105.sdi to /datadir/performance_schema/mutex_instances_105.sdi 2026-04-29T17:05:22.281844-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_105.sdi to /datadir/performance_schema/mutex_instances_105.sdi 2026-04-29T17:05:22.281877-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-04-29T17:05:22.281906-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-04-29T17:05:22.281940-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_90.sdi to /datadir/performance_schema/cond_instances_90.sdi 2026-04-29T17:05:22.281968-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_90.sdi to /datadir/performance_schema/cond_instances_90.sdi 2026-04-29T17:05:22.282001-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_162.sdi to /datadir/performance_schema/keyring_keys_162.sdi 2026-04-29T17:05:22.282025-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_162.sdi to /datadir/performance_schema/keyring_keys_162.sdi 2026-04-29T17:05:22.282054-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_197.sdi to /datadir/performance_schema/persisted_variab_197.sdi 2026-04-29T17:05:22.282080-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_197.sdi to /datadir/performance_schema/persisted_variab_197.sdi 2026-04-29T17:05:22.282115-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-04-29T17:05:22.282144-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-04-29T17:05:22.282171-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-04-29T17:05:22.282199-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-04-29T17:05:22.282229-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_189.sdi to /datadir/performance_schema/status_by_thread_189.sdi 2026-04-29T17:05:22.282257-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_189.sdi to /datadir/performance_schema/status_by_thread_189.sdi 2026-04-29T17:05:22.282286-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-04-29T17:05:22.282313-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-04-29T17:05:22.282361-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-04-29T17:05:22.282386-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-04-29T17:05:22.282414-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_188.sdi to /datadir/performance_schema/status_by_host_188.sdi 2026-04-29T17:05:22.282439-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_188.sdi to /datadir/performance_schema/status_by_host_188.sdi 2026-04-29T17:05:22.282468-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_metrics_114.sdi to /datadir/performance_schema/setup_metrics_114.sdi 2026-04-29T17:05:22.282494-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_metrics_114.sdi to /datadir/performance_schema/setup_metrics_114.sdi 2026-04-29T17:05:22.282522-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_115.sdi to /datadir/performance_schema/setup_objects_115.sdi 2026-04-29T17:05:22.282547-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_115.sdi to /datadir/performance_schema/setup_objects_115.sdi 2026-04-29T17:05:22.282579-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__161.sdi to /datadir/performance_schema/session_account__161.sdi 2026-04-29T17:05:22.282605-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__161.sdi to /datadir/performance_schema/session_account__161.sdi 2026-04-29T17:05:22.282636-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-04-29T17:05:22.282654-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-04-29T17:05:22.282682-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/account_failed_l_204.sdi to /datadir/performance_schema/account_failed_l_204.sdi 2026-04-29T17:05:22.282707-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/account_failed_l_204.sdi to /datadir/performance_schema/account_failed_l_204.sdi 2026-04-29T17:05:22.282733-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_116.sdi to /datadir/performance_schema/setup_threads_116.sdi 2026-04-29T17:05:22.282760-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_116.sdi to /datadir/performance_schema/setup_threads_116.sdi 2026-04-29T17:05:22.282788-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_171.sdi to /datadir/performance_schema/data_lock_waits_171.sdi 2026-04-29T17:05:22.282813-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_171.sdi to /datadir/performance_schema/data_lock_waits_171.sdi 2026-04-29T17:05:22.282843-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-04-29T17:05:22.282869-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-04-29T17:05:22.282896-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_110.sdi to /datadir/performance_schema/setup_actors_110.sdi 2026-04-29T17:05:22.282921-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_110.sdi to /datadir/performance_schema/setup_actors_110.sdi 2026-04-29T17:05:22.282946-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_121.sdi to /datadir/performance_schema/events_stages_cu_121.sdi 2026-04-29T17:05:22.282973-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_121.sdi to /datadir/performance_schema/events_stages_cu_121.sdi 2026-04-29T17:05:22.283002-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_169.sdi to /datadir/performance_schema/metadata_locks_169.sdi 2026-04-29T17:05:22.283027-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_169.sdi to /datadir/performance_schema/metadata_locks_169.sdi 2026-04-29T17:05:22.283054-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_159.sdi to /datadir/performance_schema/socket_summary_b_159.sdi 2026-04-29T17:05:22.283080-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_159.sdi to /datadir/performance_schema/socket_summary_b_159.sdi 2026-04-29T17:05:22.283119-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_104.sdi to /datadir/performance_schema/host_cache_104.sdi 2026-04-29T17:05:22.283148-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_104.sdi to /datadir/performance_schema/host_cache_104.sdi 2026-04-29T17:05:22.283179-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_183.sdi to /datadir/performance_schema/replication_asyn_183.sdi 2026-04-29T17:05:22.283208-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_183.sdi to /datadir/performance_schema/replication_asyn_183.sdi 2026-04-29T17:05:22.283238-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_205.sdi to /datadir/performance_schema/pxc_cluster_view_205.sdi 2026-04-29T17:05:22.283264-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_205.sdi to /datadir/performance_schema/pxc_cluster_view_205.sdi 2026-04-29T17:05:22.283294-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_95.sdi to /datadir/performance_schema/events_waits_sum_95.sdi 2026-04-29T17:05:22.283330-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_95.sdi to /datadir/performance_schema/events_waits_sum_95.sdi 2026-04-29T17:05:22.283358-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_154.sdi to /datadir/performance_schema/users_154.sdi 2026-04-29T17:05:22.283383-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_154.sdi to /datadir/performance_schema/users_154.sdi 2026-04-29T17:05:22.283412-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-04-29T17:05:22.283437-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-04-29T17:05:22.283463-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_170.sdi to /datadir/performance_schema/data_locks_170.sdi 2026-04-29T17:05:22.283487-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_170.sdi to /datadir/performance_schema/data_locks_170.sdi 2026-04-29T17:05:22.283509-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_167.sdi to /datadir/performance_schema/memory_summary_b_167.sdi 2026-04-29T17:05:22.283535-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_167.sdi to /datadir/performance_schema/memory_summary_b_167.sdi 2026-04-29T17:05:22.283564-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_195.sdi to /datadir/performance_schema/session_variable_195.sdi 2026-04-29T17:05:22.283592-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_195.sdi to /datadir/performance_schema/session_variable_195.sdi 2026-04-29T17:05:22.283622-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-04-29T17:05:22.283646-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-04-29T17:05:22.283669-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-04-29T17:05:22.283686-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-04-29T17:05:22.283709-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_194.sdi to /datadir/performance_schema/global_variables_194.sdi 2026-04-29T17:05:22.283737-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_194.sdi to /datadir/performance_schema/global_variables_194.sdi 2026-04-29T17:05:22.283769-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-04-29T17:05:22.283799-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-04-29T17:05:22.283828-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_108.sdi to /datadir/performance_schema/processlist_108.sdi 2026-04-29T17:05:22.283850-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_108.sdi to /datadir/performance_schema/processlist_108.sdi 2026-04-29T17:05:22.283880-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_120.sdi to /datadir/performance_schema/threads_120.sdi 2026-04-29T17:05:22.283904-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_120.sdi to /datadir/performance_schema/threads_120.sdi 2026-04-29T17:05:22.283936-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_123.sdi to /datadir/performance_schema/events_stages_hi_123.sdi 2026-04-29T17:05:22.283960-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_123.sdi to /datadir/performance_schema/events_stages_hi_123.sdi 2026-04-29T17:05:22.283991-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__160.sdi to /datadir/performance_schema/session_connect__160.sdi 2026-04-29T17:05:22.284021-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__160.sdi to /datadir/performance_schema/session_connect__160.sdi 2026-04-29T17:05:22.284062-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-04-29T17:05:22.284097-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-04-29T17:05:22.284128-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_128.sdi to /datadir/performance_schema/events_stages_su_128.sdi 2026-04-29T17:05:22.284154-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_128.sdi to /datadir/performance_schema/events_stages_su_128.sdi 2026-04-29T17:05:22.284182-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_92.sdi to /datadir/performance_schema/events_waits_cur_92.sdi 2026-04-29T17:05:22.284207-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_92.sdi to /datadir/performance_schema/events_waits_cur_92.sdi 2026-04-29T17:05:22.284237-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_111.sdi to /datadir/performance_schema/setup_consumers_111.sdi 2026-04-29T17:05:22.284263-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_111.sdi to /datadir/performance_schema/setup_consumers_111.sdi 2026-04-29T17:05:22.284291-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_181.sdi to /datadir/performance_schema/replication_appl_181.sdi 2026-04-29T17:05:22.284315-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_181.sdi to /datadir/performance_schema/replication_appl_181.sdi 2026-04-29T17:05:22.284357-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_119.sdi to /datadir/performance_schema/table_lock_waits_119.sdi 2026-04-29T17:05:22.284382-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_119.sdi to /datadir/performance_schema/table_lock_waits_119.sdi 2026-04-29T17:05:22.284410-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_199.sdi to /datadir/performance_schema/binary_log_trans_199.sdi 2026-04-29T17:05:22.284439-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_199.sdi to /datadir/performance_schema/binary_log_trans_199.sdi 2026-04-29T17:05:22.284469-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-04-29T17:05:22.284496-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-04-29T17:05:22.284526-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-04-29T17:05:22.284556-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-04-29T17:05:22.284584-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_190.sdi to /datadir/performance_schema/status_by_user_190.sdi 2026-04-29T17:05:22.284610-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_190.sdi to /datadir/performance_schema/status_by_user_190.sdi 2026-04-29T17:05:22.284636-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_109.sdi to /datadir/performance_schema/rwlock_instances_109.sdi 2026-04-29T17:05:22.284662-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_109.sdi to /datadir/performance_schema/rwlock_instances_109.sdi 2026-04-29T17:05:22.284692-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_196.sdi to /datadir/performance_schema/variables_info_196.sdi 2026-04-29T17:05:22.284717-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_196.sdi to /datadir/performance_schema/variables_info_196.sdi 2026-04-29T17:05:22.284747-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__106.sdi to /datadir/performance_schema/objects_summary__106.sdi 2026-04-29T17:05:22.284775-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__106.sdi to /datadir/performance_schema/objects_summary__106.sdi 2026-04-29T17:05:22.284803-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_192.sdi to /datadir/performance_schema/session_status_192.sdi 2026-04-29T17:05:22.284831-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_192.sdi to /datadir/performance_schema/session_status_192.sdi 2026-04-29T17:05:22.284860-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-04-29T17:05:22.284887-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-04-29T17:05:22.284916-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-04-29T17:05:22.284940-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-04-29T17:05:22.284970-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-04-29T17:05:22.284996-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-04-29T17:05:22.285022-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-04-29T17:05:22.285050-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-04-29T17:05:22.285078-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_186.sdi to /datadir/performance_schema/user_variables_b_186.sdi 2026-04-29T17:05:22.285108-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_186.sdi to /datadir/performance_schema/user_variables_b_186.sdi 2026-04-29T17:05:22.285137-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_140.sdi to /datadir/performance_schema/events_statement_140.sdi 2026-04-29T17:05:22.285163-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_140.sdi to /datadir/performance_schema/events_statement_140.sdi 2026-04-29T17:05:22.285193-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_155.sdi to /datadir/performance_schema/accounts_155.sdi 2026-04-29T17:05:22.285219-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_155.sdi to /datadir/performance_schema/accounts_155.sdi 2026-04-29T17:05:22.285248-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_203.sdi to /datadir/performance_schema/keyring_componen_203.sdi 2026-04-29T17:05:22.285275-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_203.sdi to /datadir/performance_schema/keyring_componen_203.sdi 2026-04-29T17:05:22.285296-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-04-29T17:05:22.285315-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-04-29T17:05:22.285383-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-04-29T17:05:22.285415-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-04-29T17:05:22.285444-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__102.sdi to /datadir/performance_schema/file_summary_by__102.sdi 2026-04-29T17:05:22.285472-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__102.sdi to /datadir/performance_schema/file_summary_by__102.sdi 2026-04-29T17:05:22.285497-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_93.sdi to /datadir/performance_schema/events_waits_his_93.sdi 2026-04-29T17:05:22.285523-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_93.sdi to /datadir/performance_schema/events_waits_his_93.sdi 2026-04-29T17:05:22.285552-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_184.sdi to /datadir/performance_schema/log_status_184.sdi 2026-04-29T17:05:22.285577-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_184.sdi to /datadir/performance_schema/log_status_184.sdi 2026-04-29T17:05:22.285605-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_200.sdi to /datadir/performance_schema/tls_channel_stat_200.sdi 2026-04-29T17:05:22.285635-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_200.sdi to /datadir/performance_schema/tls_channel_stat_200.sdi 2026-04-29T17:05:22.285665-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-04-29T17:05:22.285694-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-04-29T17:05:22.285724-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-04-29T17:05:22.285752-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-04-29T17:05:22.285784-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-04-29T17:05:22.285811-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-04-29T17:05:22.285841-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-04-29T17:05:22.285871-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-04-29T17:05:22.285903-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-04-29T17:05:22.285925-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-04-29T17:05:22.285945-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_185.sdi to /datadir/performance_schema/prepared_stateme_185.sdi 2026-04-29T17:05:22.285971-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_185.sdi to /datadir/performance_schema/prepared_stateme_185.sdi 2026-04-29T17:05:22.286001-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-04-29T17:05:22.286030-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-04-29T17:05:22.286062-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_91.sdi to /datadir/performance_schema/error_log_91.sdi 2026-04-29T17:05:22.286097-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_91.sdi to /datadir/performance_schema/error_log_91.sdi 2026-04-29T17:05:22.286126-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-04-29T17:05:22.286151-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-04-29T17:05:22.286183-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_118.sdi to /datadir/performance_schema/table_io_waits_s_118.sdi 2026-04-29T17:05:22.286211-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_118.sdi to /datadir/performance_schema/table_io_waits_s_118.sdi 2026-04-29T17:05:22.286235-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-04-29T17:05:22.286258-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-04-29T17:05:22.286290-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-04-29T17:05:22.286330-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-04-29T17:05:22.286366-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_191.sdi to /datadir/performance_schema/global_status_191.sdi 2026-04-29T17:05:22.286396-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_191.sdi to /datadir/performance_schema/global_status_191.sdi 2026-04-29T17:05:22.286427-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_157.sdi to /datadir/performance_schema/socket_instances_157.sdi 2026-04-29T17:05:22.286459-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_157.sdi to /datadir/performance_schema/socket_instances_157.sdi 2026-04-29T17:05:22.286491-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_202.sdi to /datadir/performance_schema/malloc_stats_202.sdi 2026-04-29T17:05:22.286515-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_202.sdi to /datadir/performance_schema/malloc_stats_202.sdi 2026-04-29T17:05:22.286542-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_193.sdi to /datadir/performance_schema/variables_by_thr_193.sdi 2026-04-29T17:05:22.286569-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_193.sdi to /datadir/performance_schema/variables_by_thr_193.sdi 2026-04-29T17:05:22.286608-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_174.sdi to /datadir/performance_schema/replication_conn_174.sdi 2026-04-29T17:05:22.286633-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_174.sdi to /datadir/performance_schema/replication_conn_174.sdi 2026-04-29T17:05:22.286663-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_172.sdi to /datadir/performance_schema/replication_conn_172.sdi 2026-04-29T17:05:22.286691-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_172.sdi to /datadir/performance_schema/replication_conn_172.sdi 2026-04-29T17:05:22.286720-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-04-29T17:05:22.286745-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-04-29T17:05:22.286772-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_112.sdi to /datadir/performance_schema/setup_instrument_112.sdi 2026-04-29T17:05:22.286799-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_112.sdi to /datadir/performance_schema/setup_instrument_112.sdi 2026-04-29T17:05:22.286874-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-04-29T17:05:22.286900-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-04-29T17:05:22.286977-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-04-29T17:05:22.287005-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-04-29T17:05:22.287032-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2026-04-29T17:05:22.287057-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2026-04-29T17:05:22.380253-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.b9MIoA367V Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) + rm /tmp/tmp.BKrAJL9RyN /tmp/tmp.b9MIoA367V + return 0 + wait_for_running pitr-proxysql 2 + local name=pitr-proxysql + let last_pod=1 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 1 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ echo pitr-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/pitr-proxysql-0 condition met waiting for pod/pitr-proxysql-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-1 480 + local pod=pitr-proxysql-1 + local max_retry=480 + local ns= ++ echo pitr-proxysql-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/pitr-proxysql-1 condition met waiting for pod/pitr-proxysql-1 to become Ready.Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ echo pitr-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-1 + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fgCkE62wHX +++ mktemp ++ local LAST_ERR=/tmp/tmp.1VAwDNdc3l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fgCkE62wHX ++ cat /tmp/tmp.1VAwDNdc3l ++ rm /tmp/tmp.fgCkE62wHX /tmp/tmp.1VAwDNdc3l ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dHhyDWuFVA +++ mktemp ++ local LAST_ERR=/tmp/tmp.f0Jp8tEc47 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dHhyDWuFVA ++ cat /tmp/tmp.f0Jp8tEc47 ++ rm /tmp/tmp.dHhyDWuFVA /tmp/tmp.f0Jp8tEc47 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.cCwKw6xFr8 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.gFcyMknVwa +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.cCwKw6xFr8 +++++ cat /tmp/tmp.gFcyMknVwa +++++ rm /tmp/tmp.cCwKw6xFr8 /tmp/tmp.gFcyMknVwa +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.1GV8uOw8Tp ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.rvsJManwLN +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.1GV8uOw8Tp +++++ cat /tmp/tmp.rvsJManwLN +++++ rm /tmp/tmp.1GV8uOw8Tp /tmp/tmp.rvsJManwLN +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Teiha0ELnF +++ mktemp ++ local LAST_ERR=/tmp/tmp.eR0awJnznm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Teiha0ELnF ++ cat /tmp/tmp.eR0awJnznm ++ rm /tmp/tmp.Teiha0ELnF /tmp/tmp.eR0awJnznm ++ return 0 + [[ 2 == \2 ]] + echo + desc 'check data after backup' restore-on-pitr-minio-gtid + set +o xtrace ----------------------------------------------------------------------------------- check data after backup restore-on-pitr-minio-gtid ----------------------------------------------------------------------------------- + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v2FVU8ef9B +++ mktemp ++ local LAST_ERR=/tmp/tmp.SfnYtqzkl1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v2FVU8ef9B ++ cat /tmp/tmp.SfnYtqzkl1 ++ rm /tmp/tmp.v2FVU8ef9B /tmp/tmp.SfnYtqzkl1 ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.4FyEIG03fJ/select-2.sql + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GXh7WvrtFw +++ mktemp ++ local LAST_ERR=/tmp/tmp.hYYibM1ml7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GXh7WvrtFw ++ cat /tmp/tmp.hYYibM1ml7 ++ rm /tmp/tmp.GXh7WvrtFw /tmp/tmp.hYYibM1ml7 ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-6jk4n ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.4FyEIG03fJ/select-2.sql + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LjciANUAG8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PPUmljFbpk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LjciANUAG8 ++ cat /tmp/tmp.PPUmljFbpk ++ rm /tmp/tmp.LjciANUAG8 /tmp/tmp.PPUmljFbpk ++ return 0 + client_pod=pxc-client-56fd5498cd-6jk4n + wait_pod pxc-client-56fd5498cd-6jk4n + local pod=pxc-client-56fd5498cd-6jk4n + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-6jk4n + local container= + set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.4FyEIG03fJ/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.4FyEIG03fJ/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.4FyEIG03fJ/select-2.sql + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/restore-on-pitr-minio-gtid.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.idWdyky1YW ++ mktemp + local LAST_ERR=/tmp/tmp.R2TczlHNid + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/restore-on-pitr-minio-gtid.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.idWdyky1YW perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio-gtid" deleted from pitr-31942 namespace + cat /tmp/tmp.R2TczlHNid + rm /tmp/tmp.idWdyky1YW /tmp/tmp.R2TczlHNid + return 0 ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-31942 ++ local host=pitr-pxc-0.pitr-pxc.pitr-31942 +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.qK1Slpw5Pn +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.zL0eGdLnsJ ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.qK1Slpw5Pn ++++ cat /tmp/tmp.zL0eGdLnsJ ++++ rm /tmp/tmp.qK1Slpw5Pn /tmp/tmp.zL0eGdLnsJ ++++ return 0 +++ client_pod=pxc-client-56fd5498cd-6jk4n +++ wait_pod pxc-client-56fd5498cd-6jk4n +++ local pod=pxc-client-56fd5498cd-6jk4n +++ local max_retry=480 +++ local ns= ++++ echo pxc-client-56fd5498cd-6jk4n ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ grep -E '^(pxc|proxysql)$' +++ local container= +++ set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local 'gtid=ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' ++ [[ ! ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo 'ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' + gtid='ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' + log '[CASE 1]: gtid_executed (after restore): ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:11:03+0000]' '[CASE' '1]:' gtid_executed '(after' 'restore):' 'ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' [2026-04-29T17:11:03+0000] [CASE 1]: gtid_executed (after restore): ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25 + desc '[CASE 1] PiTR with GTID: OK' + set +o xtrace ----------------------------------------------------------------------------------- [CASE 1] PiTR with GTID: OK ----------------------------------------------------------------------------------- + run_backup pitr on-pitr-minio on-pitr-minio-2 + local cluster=pitr + local backup=on-pitr-minio + local name=on-pitr-minio-2 + '[' -z on-pitr-minio-2 ']' + log 'run backup pxc-backup/on-pitr-minio-2' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:11:03+0000]' run backup pxc-backup/on-pitr-minio-2 [2026-04-29T17:11:03+0000] run backup pxc-backup/on-pitr-minio-2 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/on-pitr-minio.yml + kubectl_bin apply -f - + yq eval '.metadata.name="on-pitr-minio-2"' ++ mktemp + local LAST_OUT=/tmp/tmp.ed3RAfcRpZ ++ mktemp + local LAST_ERR=/tmp/tmp.ICKHzoitpK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ed3RAfcRpZ perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio-2 created + cat /tmp/tmp.ICKHzoitpK + rm /tmp/tmp.ed3RAfcRpZ /tmp/tmp.ICKHzoitpK + return 0 + wait_backup on-pitr-minio-2 + local backup=on-pitr-minio-2 + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio-2 to reach Succeeded state.................Succeeded + sleep_with_log 65 + local d=65 + log 'sleeping for 65 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:11:43+0000]' sleeping for 65 seconds [2026-04-29T17:11:43+0000] sleeping for 65 seconds + sleep 65 + check_binlog_gap on-pitr-minio-2 + local backup=on-pitr-minio-2 ++ grep -c 'Binlog with GTID set' ++ kubectl get pxc-backup on-pitr-minio-2 -o 'jsonpath={.status.conditions}' + [[ 0 -eq 1 ]] + desc 'no binlog gap after pxc-backup/on-pitr-minio-2: OK' + set +o xtrace ----------------------------------------------------------------------------------- no binlog gap after pxc-backup/on-pitr-minio-2: OK ----------------------------------------------------------------------------------- ++ run_mysql 'SELECT now();' '-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' ++ local 'command=SELECT now();' ++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' +++ get_client_pod +++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.poLUbxltw7 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.n3NMK3Eo2b +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.poLUbxltw7 +++ cat /tmp/tmp.n3NMK3Eo2b +++ rm /tmp/tmp.poLUbxltw7 /tmp/tmp.n3NMK3Eo2b +++ return 0 ++ client_pod=pxc-client-56fd5498cd-6jk4n ++ wait_pod pxc-client-56fd5498cd-6jk4n ++ local pod=pxc-client-56fd5498cd-6jk4n ++ local max_retry=480 ++ local ns= +++ grep -E '^(pxc|proxysql)$' +++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' +++ echo pxc-client-56fd5498cd-6jk4n ++ local container= ++ set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ++ set +o xtrace + time_now='2026-04-29 17:13:05' ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-31942 ++ local host=pitr-pxc-0.pitr-pxc.pitr-31942 +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-31942 -uroot -proot_password' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.COJgF3nh52 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.7E7WYvZ1D8 ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.COJgF3nh52 ++++ cat /tmp/tmp.7E7WYvZ1D8 ++++ rm /tmp/tmp.COJgF3nh52 /tmp/tmp.7E7WYvZ1D8 ++++ return 0 +++ client_pod=pxc-client-56fd5498cd-6jk4n +++ wait_pod pxc-client-56fd5498cd-6jk4n +++ local pod=pxc-client-56fd5498cd-6jk4n +++ local max_retry=480 +++ local ns= ++++ grep -E '^(pxc|proxysql)$' ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ echo pxc-client-56fd5498cd-6jk4n +++ local container= +++ set +o xtrace pod/pxc-client-56fd5498cd-6jk4n condition met waiting for pod/pxc-client-56fd5498cd-6jk4n to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local 'gtid=ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' ++ [[ ! ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo 'ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' + gtid='ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' + log '[CASE 2]: gtid_executed: ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:13:21+0000]' '[CASE' '2]:' gtid_executed: 'ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25' [2026-04-29T17:13:21+0000] [CASE 2]: gtid_executed: ac9e3618-43ed-11f1-bb51-ea3cc3537194:1,\ne3794f29-43ea-11f1-a5b1-32c5d2a852bf:25 + write_data_for_pitr pitr 100503 3 + local cluster=pitr + local start=100503 + local rows=3 ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mJrhvngU5Q ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Rb0PfM5kpd +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.mJrhvngU5Q +++ cat /tmp/tmp.Rb0PfM5kpd +++ rm /tmp/tmp.mJrhvngU5Q /tmp/tmp.Rb0PfM5kpd +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6EzO9IVjO6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bxSZFa8Yns +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.6EzO9IVjO6 +++ cat /tmp/tmp.bxSZFa8Yns +++ rm /tmp/tmp.6EzO9IVjO6 /tmp/tmp.bxSZFa8Yns +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + local sql= + (( i=0 )) + (( i//g' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-31942~ + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/restore-on-pitr-minio-time.yaml + /usr/bin/sed -e 's///g' + /usr/bin/sed -e 's//2026-04-29 17:13:05/g' ++ mktemp + local LAST_OUT=/tmp/tmp.tpxpJgRXDI ++ mktemp + local LAST_ERR=/tmp/tmp.y8I2Iih3v5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tpxpJgRXDI perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-time created + cat /tmp/tmp.y8I2Iih3v5 + rm /tmp/tmp.tpxpJgRXDI /tmp/tmp.y8I2Iih3v5 + return 0 + wait_backup_restore restore-on-pitr-minio-time 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-time + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-time to reach Stopping Cluster state 2026-04-29T17:13:44 pxc-restore/restore-on-pitr-minio-time state: Stopping Cluster + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace .Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ...Error from server (NotFound): pods "pitr-pxc-1" not found + wait_for_delete pod/pitr-pxc-0 + local res=pod/pitr-pxc-0 + echo -n 'waiting for pod/pitr-pxc-0 to be deleted' waiting for pod/pitr-pxc-0 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-0" not found + wait_backup_restore restore-on-pitr-minio-time 'Point-in-time recovering' + local backup_name=restore-on-pitr-minio-time + local 'target_state=Point-in-time recovering' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-time to reach Point-in-time recovering state 2026-04-29T17:14:25 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:29 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:33 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:37 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:41 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:43 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:46 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:49 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:52 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:55 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:14:58 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:00 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:03 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:06 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:10 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:13 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:16 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:19 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:22 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:26 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:29 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:32 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:35 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:39 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:41 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:44 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:15:46 pxc-restore/restore-on-pitr-minio-time state: Point-in-time recovering + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ........................................................................................................................2026-04-29T16:40:41.455Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.32.13-gke.1362000"} 2026-04-29T16:40:41.455Z INFO setup Manager starting up {"gitCommit": "9bec550e38b53ea8a42441eabf05aff8b718c288", "gitBranch": "PR-2433-9bec550e", "buildTime": "2026-04-29T14:21:52Z", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} 2026-04-29T16:40:41.456Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "", "enabled": ""} 2026-04-29T16:40:41.459Z INFO setup Registering Components. 2026-04-29T16:40:42.038Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-04-29T16:40:42.038Z INFO setup Starting the Cmd. 2026-04-29T16:40:42.038Z INFO controller-runtime.metrics Starting metrics server 2026-04-29T16:40:42.038Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-04-29T16:40:42.038Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-04-29T16:40:42.038Z INFO controller-runtime.webhook Starting webhook server 2026-04-29T16:40:42.039Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-04-29T16:40:42.039Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-04-29T16:40:42.084Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-04-29T16:40:42.140Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:40:42.167Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:40:42.167Z DEBUG events percona-xtradb-cluster-operator-7cd7bb76f4-kwnbc_2c316b0a-dcf2-4a95-a1d1-41a219a46e24 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"674dc412-bc91-4b3b-a8cc-f71091bd4025","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1777480842161295009"}, "reason": "LeaderElection"} 2026-04-29T16:40:42.167Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-04-29T16:40:42.167Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-04-29T16:40:42.167Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-04-29T16:40:42.167Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-04-29T16:40:42.268Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-04-29T16:40:42.268Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-04-29T16:40:42.268Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-04-29T16:40:42.268Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-04-29T16:40:42.368Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-04-29T16:40:42.368Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-04-29T16:45:23.928Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "version": "1.20.0"} 2026-04-29T16:45:24.227Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e"} 2026-04-29T16:45:27.270Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e"} 2026-04-29T16:45:27.294Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e"} 2026-04-29T16:45:30.494Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "auto-pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:45:30.518Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:45:30.642Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:45:30.708Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:45:30.820Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:45:30.893Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "pitr-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:45:31.007Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:45:31.098Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ff3f560c-f02a-4f45-b028-e232d0ae9c2e", "object": "pitr-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:45:32.310Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "435d6318-1666-4582-9143-1faa08868a23", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777481131045007007", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiNjBkNzcxY2M2MGE0NGVhOGY5NTYzYmJlMDJhYjUyNTQifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InRtcCIsImVtcHR5RGlyIjp7fX0seyJuYW1lIjoiY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJwaXRyLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsLWludGVybmFsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbC1pbnRlcm5hbCIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbCIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6ImF1dG8tY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJhdXRvLXBpdHItcHhjIiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJ2YXVsdC1rZXlyaW5nLXNlY3JldCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoic29tZS1uYW1lLXZhdWx0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJteXNxbC11c2Vycy1zZWNyZXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoiaW50ZXJuYWwtcGl0ciIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6Im15c3FsLWluaXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1teXNxbC1pbml0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJkYXRhZGlyIiwibW91bnRQYXRoIjoiL3Zhci9saWIvbXlzcWwifSx7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvb3B0L3BlcmNvbmEifV0sImltYWdlUHVsbFBvbGljeSI6IkFsd2F5cyJ9XSwiY29udGFpbmVycyI6W3sibmFtZSI6InB4YyIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOm1haW4tcHhjOC40IiwiY29tbWFuZCI6WyIvdmFyL2xpYi9teXNxbC9weGMtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbIm15c3FsZCJdLCJwb3J0cyI6W3sibmFtZSI6Im15c3FsIiwiY29udGFpbmVyUG9ydCI6MzMwNn0seyJuYW1lIjoic3N0IiwiY29udGFpbmVyUG9ydCI6NDQ0NH0seyJuYW1lIjoid3JpdGUtc2V0IiwiY29udGFpbmVyUG9ydCI6NDU2N30seyJuYW1lIjoiaXN0IiwiY29udGFpbmVyUG9ydCI6NDU2OH0seyJuYW1lIjoibXlzcWwtYWRtaW4iLCJjb250YWluZXJQb3J0IjozMzA2Mn0seyJuYW1lIjoibXlzcWx4IiwiY29udGFpbmVyUG9ydCI6MzMwNjB9XSwiZW52RnJvbSI6W3sic2VjcmV0UmVmIjp7Im5hbWUiOiJwaXRyLWVudi12YXJzLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMtdW5yZWFkeSJ9LHsibmFtZSI6Ik1PTklUT1JfSE9TVCIsInZhbHVlIjoiJSJ9LHsibmFtZSI6Ik1ZU1FMX1JPT1RfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5Ijoicm9vdCJ9fX0seyJuYW1lIjoiWFRSQUJBQ0tVUF9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJ4dHJhYmFja3VwIn19fSx7Im5hbWUiOiJNT05JVE9SX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im1vbml0b3IifX19LHsibmFtZSI6IkNMVVNURVJfSEFTSCIsInZhbHVlIjoiMTg5NTg5OSJ9LHsibmFtZSI6Ik9QRVJBVE9SX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im9wZXJhdG9yIn19fSx7Im5hbWUiOiJMSVZFTkVTU19DSEVDS19USU1FT1VUIiwidmFsdWUiOiI1In0seyJuYW1lIjoiUkVBRElORVNTX0NIRUNLX1RJTUVPVVQiLCJ2YWx1ZSI6IjE1In0seyJuYW1lIjoiREVGQVVMVF9BVVRIRU5USUNBVElPTl9QTFVHSU4iLCJ2YWx1ZSI6ImNhY2hpbmdfc2hhMl9wYXNzd29yZCJ9LHsibmFtZSI6Ik1ZU1FMX05PVElGWV9TT0NLRVQiLCJ2YWx1ZSI6Ii92YXIvbGliL215c3FsL25vdGlmeS5zb2NrIn0seyJuYW1lIjoiTVlTUUxfU1RBVEVfRklMRSIsInZhbHVl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMmFiMWU5YzdhMThhYzM4ZDJhZWE4MjRmZDY0YWIxMjcifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InRtcCIsImVtcHR5RGlyIjp7fX0seyJuYW1lIjoiY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJwaXRyLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsLWludGVybmFsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbC1pbnRlcm5hbCIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbCIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6ImF1dG8tY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJhdXRvLXBpdHItcHhjIiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJ2YXVsdC1rZXlyaW5nLXNlY3JldCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoic29tZS1uYW1lLXZhdWx0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJteXNxbC11c2Vycy1zZWNyZXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoiaW50ZXJuYWwtcGl0ciIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6Im15c3FsLWluaXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1teXNxbC1pbml0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJkYXRhZGlyIiwibW91bnRQYXRoIjoiL3Zhci9saWIvbXlzcWwifSx7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvb3B0L3BlcmNvbmEifV0sImltYWdlUHVsbFBvbGljeSI6IkFsd2F5cyJ9XSwiY29udGFpbmVycyI6W3sibmFtZSI6InB4YyIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOm1haW4tcHhjOC40IiwiY29tbWFuZCI6WyIvdmFyL2xpYi9teXNxbC9weGMtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbIm15c3FsZCJdLCJwb3J0cyI6W3sibmFtZSI6Im15c3FsIiwiY29udGFpbmVyUG9ydCI6MzMwNn0seyJuYW1lIjoic3N0IiwiY29udGFpbmVyUG9ydCI6NDQ0NH0seyJuYW1lIjoid3JpdGUtc2V0IiwiY29udGFpbmVyUG9ydCI6NDU2N30seyJuYW1lIjoiaXN0IiwiY29udGFpbmVyUG9ydCI6NDU2OH0seyJuYW1lIjoibXlzcWwtYWRtaW4iLCJjb250YWluZXJQb3J0IjozMzA2Mn0seyJuYW1lIjoibXlzcWx4IiwiY29udGFpbmVyUG9ydCI6MzMwNjB9XSwiZW52RnJvbSI6W3sic2VjcmV0UmVmIjp7Im5hbWUiOiJwaXRyLWVudi12YXJzLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMtdW5yZWFkeSJ9LHsibmFtZSI6Ik1PTklUT1JfSE9TVCIsInZhbHVlIjoiJSJ9LHsibmFtZSI6Ik1ZU1FMX1JPT1RfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5Ijoicm9vdCJ9fX0seyJuYW1lIjoiWFRSQUJBQ0tVUF9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJ4dHJhYmFja3VwIn19fSx7Im5hbWUiOiJNT05JVE9SX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im1vbml0b3IifX19LHsibmFtZSI6IkNMVVNURVJfSEFTSCIsInZhbHVlIjoiMTg5NTg5OSJ9LHsibmFtZSI6Ik9QRVJBVE9SX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im9wZXJhdG9yIn19fSx7Im5hbWUiOiJMSVZFTkVTU19DSEVDS19USU1FT1VUIiwidmFsdWUiOiI1In0seyJuYW1lIjoiUkVBRElORVNTX0NIRUNLX1RJTUVPVVQiLCJ2YWx1ZSI6IjE1In0seyJuYW1lIjoiREVGQVVMVF9BVVRIRU5USUNBVElPTl9QTFVHSU4iLCJ2YWx1ZSI6ImNhY2hpbmdfc2hhMl9wYXNzd29yZCJ9LHsibmFtZSI6Ik1ZU1FMX05PVElGWV9TT0NLRVQiLCJ2YWx1ZSI6Ii92YXIvbGliL215c3FsL25vdGlmeS5zb2NrIn0seyJuYW1lIjoiTVlTUUxfU1RBVEVfRklMRSIsInZhbHVl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:30 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:31 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:collisionCount":{},"f:currentRevision":{},"f:obs`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": strings.Join({ -  "60d771cc60a44ea8f9563bbe02ab5254", +  "2ab1e9c7a18ac38d2aea824fd64ab127",    }, ""),    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-5f44b49d7f", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-5f44b49d7f", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:45:32.977Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "435d6318-1666-4582-9143-1faa08868a23", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:45:33.027Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "435d6318-1666-4582-9143-1faa08868a23", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777481131134655003", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiNjBkNzcxY2M2MGE0NGVhOGY5NTYzYmJlMDJhYjUyNTQifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMmFiMWU5YzdhMThhYzM4ZDJhZWE4MjRmZDY0YWIxMjcifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:30 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:31 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:collisionCount":{},"f:currentReplicas":{},"f:cur`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "proxysql",    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": strings.Join({ -  "60d771cc60a44ea8f9563bbe02ab5254", +  "2ab1e9c7a18ac38d2aea824fd64ab127",    }, ""),    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0,    ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-577f5ffff6", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-577f5ffff6", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:45:33.167Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "435d6318-1666-4582-9143-1faa08868a23", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:45:33.198Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "435d6318-1666-4582-9143-1faa08868a23", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777481131134655003", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiNjBkNzcxY2M2MGE0NGVhOGY5NTYzYmJlMDJhYjUyNTQifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMmFiMWU5YzdhMThhYzM4ZDJhZWE4MjRmZDY0YWIxMjcifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:30 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:31 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:collisionCount":{},"f:currentReplicas":{},"f:cur`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "proxysql",    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": strings.Join({ -  "60d771cc60a44ea8f9563bbe02ab5254", +  "2ab1e9c7a18ac38d2aea824fd64ab127",    }, ""),    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0,    ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-577f5ffff6", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-577f5ffff6", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:45:38.474Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "091cd560-9d0f-413f-96a1-378ba1ce2e06", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777481133268943007", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMmFiMWU5YzdhMThhYzM4ZDJhZWE4MjRmZDY0YWIxMjcifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InRtcCIsImVtcHR5RGlyIjp7fX0seyJuYW1lIjoiY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJwaXRyLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsLWludGVybmFsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbC1pbnRlcm5hbCIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbCIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6ImF1dG8tY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJhdXRvLXBpdHItcHhjIiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJ2YXVsdC1rZXlyaW5nLXNlY3JldCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoic29tZS1uYW1lLXZhdWx0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJteXNxbC11c2Vycy1zZWNyZXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoiaW50ZXJuYWwtcGl0ciIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6Im15c3FsLWluaXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1teXNxbC1pbml0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJkYXRhZGlyIiwibW91bnRQYXRoIjoiL3Zhci9saWIvbXlzcWwifSx7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvb3B0L3BlcmNvbmEifV0sImltYWdlUHVsbFBvbGljeSI6IkFsd2F5cyJ9XSwiY29udGFpbmVycyI6W3sibmFtZSI6InB4YyIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOm1haW4tcHhjOC40IiwiY29tbWFuZCI6WyIvdmFyL2xpYi9teXNxbC9weGMtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbIm15c3FsZCJdLCJwb3J0cyI6W3sibmFtZSI6Im15c3FsIiwiY29udGFpbmVyUG9ydCI6MzMwNn0seyJuYW1lIjoic3N0IiwiY29udGFpbmVyUG9ydCI6NDQ0NH0seyJuYW1lIjoid3JpdGUtc2V0IiwiY29udGFpbmVyUG9ydCI6NDU2N30seyJuYW1lIjoiaXN0IiwiY29udGFpbmVyUG9ydCI6NDU2OH0seyJuYW1lIjoibXlzcWwtYWRtaW4iLCJjb250YWluZXJQb3J0IjozMzA2Mn0seyJuYW1lIjoibXlzcWx4IiwiY29udGFpbmVyUG9ydCI6MzMwNjB9XSwiZW52RnJvbSI6W3sic2VjcmV0UmVmIjp7Im5hbWUiOiJwaXRyLWVudi12YXJzLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMtdW5yZWFkeSJ9LHsibmFtZSI6Ik1PTklUT1JfSE9TVCIsInZhbHVlIjoiJSJ9LHsibmFtZSI6Ik1ZU1FMX1JPT1RfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5Ijoicm9vdCJ9fX0seyJuYW1lIjoiWFRSQUJBQ0tVUF9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJ4dHJhYmFja3VwIn19fSx7Im5hbWUiOiJNT05JVE9SX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im1vbml0b3IifX19LHsibmFtZSI6IkNMVVNURVJfSEFTSCIsInZhbHVlIjoiMTg5NTg5OSJ9LHsibmFtZSI6Ik9QRVJBVE9SX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im9wZXJhdG9yIn19fSx7Im5hbWUiOiJMSVZFTkVTU19DSEVDS19USU1FT1VUIiwidmFsdWUiOiI1In0seyJuYW1lIjoiUkVBRElORVNTX0NIRUNLX1RJTUVPVVQiLCJ2YWx1ZSI6IjE1In0seyJuYW1lIjoiREVGQVVMVF9BVVRIRU5USUNBVElPTl9QTFVHSU4iLCJ2YWx1ZSI6ImNhY2hpbmdfc2hhMl9wYXNzd29yZCJ9LHsibmFtZSI6Ik1ZU1FMX05PVElGWV9TT0NLRVQiLCJ2YWx1ZSI6Ii92YXIvbGliL215c3FsL25vdGlmeS5zb2NrIn0seyJuYW1lIjoiTVlTUUxfU1RBVEVfRklMRSIsInZhbHVl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZDcyMTA3NjQ2NDQ2NWNjOTk5YmU3MWVlNTU4ZGE4NjkifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InRtcCIsImVtcHR5RGlyIjp7fX0seyJuYW1lIjoiY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJwaXRyLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsLWludGVybmFsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbC1pbnRlcm5hbCIsIm9wdGlvbmFsIjp0cnVlfX0seyJuYW1lIjoic3NsIiwic2VjcmV0Ijp7InNlY3JldE5hbWUiOiJwaXRyLXNzbCIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6ImF1dG8tY29uZmlnIiwiY29uZmlnTWFwIjp7Im5hbWUiOiJhdXRvLXBpdHItcHhjIiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJ2YXVsdC1rZXlyaW5nLXNlY3JldCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoic29tZS1uYW1lLXZhdWx0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJteXNxbC11c2Vycy1zZWNyZXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoiaW50ZXJuYWwtcGl0ciIsIm9wdGlvbmFsIjpmYWxzZX19LHsibmFtZSI6Im15c3FsLWluaXQtZmlsZSIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1teXNxbC1pbml0Iiwib3B0aW9uYWwiOnRydWV9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJkYXRhZGlyIiwibW91bnRQYXRoIjoiL3Zhci9saWIvbXlzcWwifSx7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvb3B0L3BlcmNvbmEifV0sImltYWdlUHVsbFBvbGljeSI6IkFsd2F5cyJ9XSwiY29udGFpbmVycyI6W3sibmFtZSI6InB4YyIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOm1haW4tcHhjOC40IiwiY29tbWFuZCI6WyIvdmFyL2xpYi9teXNxbC9weGMtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbIm15c3FsZCJdLCJwb3J0cyI6W3sibmFtZSI6Im15c3FsIiwiY29udGFpbmVyUG9ydCI6MzMwNn0seyJuYW1lIjoic3N0IiwiY29udGFpbmVyUG9ydCI6NDQ0NH0seyJuYW1lIjoid3JpdGUtc2V0IiwiY29udGFpbmVyUG9ydCI6NDU2N30seyJuYW1lIjoiaXN0IiwiY29udGFpbmVyUG9ydCI6NDU2OH0seyJuYW1lIjoibXlzcWwtYWRtaW4iLCJjb250YWluZXJQb3J0IjozMzA2Mn0seyJuYW1lIjoibXlzcWx4IiwiY29udGFpbmVyUG9ydCI6MzMwNjB9XSwiZW52RnJvbSI6W3sic2VjcmV0UmVmIjp7Im5hbWUiOiJwaXRyLWVudi12YXJzLXB4YyIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMtdW5yZWFkeSJ9LHsibmFtZSI6Ik1PTklUT1JfSE9TVCIsInZhbHVlIjoiJSJ9LHsibmFtZSI6Ik1ZU1FMX1JPT1RfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5Ijoicm9vdCJ9fX0seyJuYW1lIjoiWFRSQUJBQ0tVUF9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJ4dHJhYmFja3VwIn19fSx7Im5hbWUiOiJNT05JVE9SX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im1vbml0b3IifX19LHsibmFtZSI6IkNMVVNURVJfSEFTSCIsInZhbHVlIjoiMTg5NTg5OSJ9LHsibmFtZSI6Ik9QRVJBVE9SX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6Im9wZXJhdG9yIn19fSx7Im5hbWUiOiJMSVZFTkVTU19DSEVDS19USU1FT1VUIiwidmFsdWUiOiI1In0seyJuYW1lIjoiUkVBRElORVNTX0NIRUNLX1RJTUVPVVQiLCJ2YWx1ZSI6IjE1In0seyJuYW1lIjoiREVGQVVMVF9BVVRIRU5USUNBVElPTl9QTFVHSU4iLCJ2YWx1ZSI6ImNhY2hpbmdfc2hhMl9wYXNzd29yZCJ9LHsibmFtZSI6Ik1ZU1FMX05PVElGWV9TT0NLRVQiLCJ2YWx1ZSI6Ii92YXIvbGliL215c3FsL25vdGlmeS5zb2NrIn0seyJuYW1lIjoiTVlTUUxfU1RBVEVfRklMRSIsInZhbHVl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:32 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:33 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:collisionCount":{},"f:currentReplicas":{},"f:cur`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": strings.Join({ -  "2ab1e9c7a18ac38d2aea824fd64ab127", +  "d721076464465cc999be71ee558da869",    }, ""),    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0,    ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-5f44b49d7f", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-69f9dff5f7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:45:38.524Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "091cd560-9d0f-413f-96a1-378ba1ce2e06", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777481133383055003", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMmFiMWU5YzdhMThhYzM4ZDJhZWE4MjRmZDY0YWIxMjcifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZDcyMTA3NjQ2NDQ2NWNjOTk5YmU3MWVlNTU4ZGE4NjkifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:33 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:collisionCount":{},"f:currentReplicas":{},"f:cur`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:33 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "proxysql",    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": strings.Join({ -  "2ab1e9c7a18ac38d2aea824fd64ab127", +  "d721076464465cc999be71ee558da869",    }, ""),    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0,    ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-577f5ffff6", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-654bbbf66", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:45:38.603Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "091cd560-9d0f-413f-96a1-378ba1ce2e06", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777481133383055003", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMmFiMWU5YzdhMThhYzM4ZDJhZWE4MjRmZDY0YWIxMjcifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZDcyMTA3NjQ2NDQ2NWNjOTk5YmU3MWVlNTU4ZGE4NjkifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:33 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:collisionCount":{},"f:currentReplicas":{},"f:cur`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:33 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "proxysql",    "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": strings.Join({ -  "2ab1e9c7a18ac38d2aea824fd64ab127", +  "d721076464465cc999be71ee558da869",    }, ""),    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0,    ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-577f5ffff6", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-654bbbf66", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:49:40.029Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6a59fbf8-c4cc-4e36-8ad3-2a4ecae5fc33", "err": "failed to connect to pod pitr-pxc-0: dial tcp 10.158.161.57:33062: connect: connection refused"} 2026-04-29T16:49:45.206Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f", "user": "operator"} 2026-04-29T16:49:45.248Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f", "user": "monitor"} 2026-04-29T16:49:45.364Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f"} 2026-04-29T16:49:45.405Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f"} 2026-04-29T16:49:45.438Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f", "user": "xtrabackup"} 2026-04-29T16:49:45.513Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f"} 2026-04-29T16:49:45.552Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f", "user": "replication"} 2026-04-29T16:49:45.562Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7662f010-3270-4c51-8f2b-04ccfc4e428f", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:49:50.703Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "21a1692f-be9c-4327-877b-783b114c67d2", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:49:55.820Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "eab330b4-b1f2-4f33-83fe-abbef84d241d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:50:00.949Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "1fdd0079-6d3f-468a-9fec-fd147964dc15", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:50:06.065Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7aa6f1bd-8e9c-4d19-84f2-70e6844aec73", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:50:11.174Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f2c40e3f-355d-4a31-bc54-96905bb599c7", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:50:16.318Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "fe500306-7547-4dee-92f8-93a211d0fe8a", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:50:22.530Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8dc5e9b5-88f6-43fa-bca3-e8da16facafb", "user": "root"} 2026-04-29T16:50:22.570Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8dc5e9b5-88f6-43fa-bca3-e8da16facafb", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}"} 2026-04-29T16:50:22.630Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8dc5e9b5-88f6-43fa-bca3-e8da16facafb", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T16:50:22.683Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8dc5e9b5-88f6-43fa-bca3-e8da16facafb", "new version": "8.4.7-7.1"} 2026-04-29T16:50:23.390Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "d8921970-0840-4d24-9dca-dba8fb2037b6", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}", "hashChanged": false, "metaChanged": true}   &v1.Service{    TypeMeta: v1.TypeMeta{ -  Kind: "Service", +  Kind: "", -  APIVersion: "v1", +  APIVersion: "",    },    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5c2b9c17-9a85-4862-b850-7675f43c3db0", +  UID: "", -  ResourceVersion: "1777481422623999012", +  ResourceVersion: "",    Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:50:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil,    Annotations: map[string]string{ -  "cloud.google.com/neg": `{"ingress":true}`,    "percona.com/last-config-hash": "eyJwb3J0cyI6W3sibmFtZSI6Imh0dHAiLCJwb3J0Ijo4MDgwLCJ0YXJnZXRQb3J0"...,    },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "v1", -  Time: s"2026-04-29 16:50:22 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.ServiceSpec{    Ports: []v1.ServicePort{    {    Name: "http", -  Protocol: "TCP", +  Protocol: "",    AppProtocol: nil,    Port: 8080,    TargetPort: intstr.IntOrString{    Type: 0, -  IntVal: 8080, +  IntVal: 0,    StrVal: "",    },    NodePort: 0,    },    },    Selector: {"app.kubernetes.io/component": "pitr", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    ClusterIP: "34.118.229.79", -  ClusterIPs: []string{"34.118.229.79"}, +  ClusterIPs: nil,    Type: "ClusterIP",    ExternalIPs: nil, -  SessionAffinity: "None", +  SessionAffinity: "",    LoadBalancerIP: "",    LoadBalancerSourceRanges: nil,    ... // 3 identical fields    PublishNotReadyAddresses: false,    SessionAffinityConfig: nil, -  IPFamilies: []v1.IPFamily{"IPv4"}, +  IPFamilies: nil, -  IPFamilyPolicy: &"SingleStack", +  IPFamilyPolicy: nil,    AllocateLoadBalancerNodePorts: nil,    LoadBalancerClass: nil, -  InternalTrafficPolicy: &"Cluster", +  InternalTrafficPolicy: nil,    TrafficDistribution: nil,    },    Status: {},   } 2026-04-29T16:50:24.472Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8dc5e9b5-88f6-43fa-bca3-e8da16facafb"} 2026-04-29T16:50:30.674Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6c7f4921-2c9a-4fff-9fe3-86f38d33d56a"} 2026-04-29T16:50:35.874Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "a2c156d7-b90d-4109-b7e4-50f89b81082b"} 2026-04-29T16:50:41.441Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8ae3f77e-959c-41f0-a0ea-adeac2c1db27"} 2026-04-29T16:50:46.766Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "096039a8-2490-4309-abad-61bedc7be6dd"} 2026-04-29T16:50:52.048Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5f786f86-d09e-4fcc-8645-46ac4771ae32"} 2026-04-29T16:50:57.061Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ce3c7053-f8b6-447e-a210-cea3e4107ee8"} 2026-04-29T16:51:02.673Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "26a198c9-8346-46d1-a225-8e97ca5682a0"} 2026-04-29T16:51:07.764Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "11e2e135-94d3-41d3-969e-f58495ddc64c"} 2026-04-29T16:51:07.791Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "666679b8-3b24-410a-b4c7-8ecbe5116c33", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:07.834Z INFO Created a new backup job {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "666679b8-3b24-410a-b4c7-8ecbe5116c33", "namespace": "pitr-31942", "name": "xb-on-pitr-minio"} 2026-04-29T16:51:12.835Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "74d4c334-167c-46c6-b725-650fa043fffb", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:12.903Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "44b2ba3e-9507-46a2-84d3-2126d236f642", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:12.957Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "fde8d998-cea5-4288-a008-c5bcbdac2fb6"} 2026-04-29T16:51:17.938Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "15019fb3-4971-45ed-a8da-6aba3e47dbfb", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:18.556Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "91300793-6838-4452-be1a-88679b5209ec"} 2026-04-29T16:51:22.987Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "c09f57fc-3ed0-4e2e-9d8f-ec1a68cba859", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:23.863Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "554c9edb-a45a-4361-8888-af6721f17d40"} 2026-04-29T16:51:28.024Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "2411b586-9ab5-4eec-8e1f-91327827155d", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:29.160Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "51bf06a6-da1c-4659-a094-dbf5f35e4a2c"} 2026-04-29T16:51:33.058Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "f33d56b4-016f-473f-8ba6-bb6b295862d5", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:34.588Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "42bdd63b-8752-4b45-803a-c9e458472192"} 2026-04-29T16:51:38.121Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "0280f9aa-3f6c-4462-a5a4-0f6e52983378", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:39.871Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b131d591-a323-4ce1-be5e-15492f9f9be7"} 2026-04-29T16:51:43.162Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "b2f24813-56cf-4b38-9599-1b32d353b195", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:51:43.197Z INFO Backup succeeded {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "b2f24813-56cf-4b38-9599-1b32d353b195", "job": "xb-on-pitr-minio"} 2026-04-29T16:51:43.197Z DEBUG Removing binlog gap file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "b2f24813-56cf-4b38-9599-1b32d353b195", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-798584cd4f-jzrm4"} 2026-04-29T16:51:44.092Z DEBUG Removing binlog timeline file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "b2f24813-56cf-4b38-9599-1b32d353b195", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-798584cd4f-jzrm4"} 2026-04-29T16:51:44.390Z DEBUG Removing mysql-init secret {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio", "reconcileID": "b2f24813-56cf-4b38-9599-1b32d353b195", "job": "xb-on-pitr-minio", "secret": "pitr-mysql-init"} 2026-04-29T16:51:45.059Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e316ac16-4045-4772-bfd2-77d40b6b9923"} 2026-04-29T16:51:50.573Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e6e56843-7cf0-4ff2-a2f6-940c7025f469"} 2026-04-29T16:51:56.263Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "0bf175cd-b3f5-4152-a149-3f55abdff0b3"} 2026-04-29T16:52:01.883Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b059d9e0-1b67-41b2-9074-74fdde2facbd"} 2026-04-29T16:52:07.264Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6ec989c4-745f-40ee-bc34-cc6f67ccbb72"} 2026-04-29T16:52:13.264Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "85670833-e81b-4d38-ba35-9288e47f8f28"} 2026-04-29T16:52:19.373Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "eba4cf68-10eb-4cab-b343-812169494567"} 2026-04-29T16:52:23.648Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "140db7df-489b-4279-927e-efdb8c6f3982", "latest": "2026-04-29 16:52:20 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:52:24.664Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "140db7df-489b-4279-927e-efdb8c6f3982"} 2026-04-29T16:52:30.678Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "2776a4a0-339d-4f10-90f7-1b5aa45168f4"} 2026-04-29T16:52:36.493Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "460d8d51-0658-493e-9356-6e4a08cba8f4"} 2026-04-29T16:52:42.164Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "cc78c1e9-4d13-44d8-8e72-a12eebf3d2b0"} 2026-04-29T16:52:48.144Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "93850e5b-c0de-4317-80af-11481bf3316e"} 2026-04-29T16:52:54.234Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f2cf19cb-7411-4441-a55f-192fe6d33ec4"} 2026-04-29T16:52:59.961Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "a70dc81e-1228-4640-a03c-9e7a820ae136"} 2026-04-29T16:53:02.427Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7", "user": "xtrabackup"} 2026-04-29T16:53:02.447Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7", "user": "xtrabackup"} 2026-04-29T16:53:02.465Z INFO MySQL init secret created {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T16:53:02.487Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7", "user": "xtrabackup"} 2026-04-29T16:53:02.502Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7", "user": "xtrabackup"} 2026-04-29T16:53:02.506Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7", "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258"} 2026-04-29T16:53:02.509Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777481419337695007", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:38 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:50:19 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc", +  "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258",    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6d865c8f5d", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6d865c8f5d", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:53:04.881Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e0891e6-66e3-477c-a2e2-bf0b3edd90e7"} 2026-04-29T16:53:44.477Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "04d79147-923f-4c32-b6ab-2647d6c4471b", "latest": "2026-04-29 16:53:40 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:54:07.086Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8a3ce09b-328f-44a3-91c8-668ed0afef39", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T16:54:12.723Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c56f8d85-e455-4262-b5cc-4dd51deb6a7e", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T16:55:09.136Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "aed1f9f8-0dc1-4e98-8b39-ecc4acf72898", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:14.864Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "3216f515-9c1b-4898-90b9-dc6a48c58415", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:20.322Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "006c8094-c207-40f0-b088-10c89a52ec70", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:25.990Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "2a258188-d753-4d1c-b445-0026abbdd965", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:31.545Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "a898d3c7-a459-47b7-9ca5-210a0518b8f7", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:37.503Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "08ff34fe-7ff8-4e60-bf89-9a45941f29c0", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:38.123Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "08ff34fe-7ff8-4e60-bf89-9a45941f29c0", "latest": "2026-04-29 16:54:16 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:55:43.238Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "cb0fb02c-d003-4db4-918b-1db234a474f3", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:48.735Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "bd20db4e-1759-493d-8ec1-d0e102b74752", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:55:58.292Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "54d93b73-c92d-4824-8e26-0fcd4a24e1ec"} 2026-04-29T16:56:02.945Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c596fbd1-973e-4e51-9959-354ec4f1f1d3", "user": "xtrabackup"} 2026-04-29T16:56:02.966Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c596fbd1-973e-4e51-9959-354ec4f1f1d3", "user": "xtrabackup"} 2026-04-29T16:56:02.987Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c596fbd1-973e-4e51-9959-354ec4f1f1d3", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T16:56:03.007Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c596fbd1-973e-4e51-9959-354ec4f1f1d3", "user": "xtrabackup"} 2026-04-29T16:56:03.024Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c596fbd1-973e-4e51-9959-354ec4f1f1d3", "user": "xtrabackup"} 2026-04-29T16:56:03.027Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c596fbd1-973e-4e51-9959-354ec4f1f1d3", "last-applied-secret": "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d"} 2026-04-29T16:56:03.030Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c596fbd1-973e-4e51-9959-354ec4f1f1d3", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777481749821135007", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:53:02 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:55:49 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258", +  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-7867f8cbdf", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-7867f8cbdf", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:56:04.123Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7ee8ce7c-fba0-4b69-b21d-52022091279c"} 2026-04-29T16:56:42.491Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "77e070ab-99ed-47ed-918a-3b574627546a", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:56:44.753Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "434f786c-c4d5-49ee-a3b3-0505d1100cde", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:56:47.532Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "abd96d18-4329-4a55-8804-26f2be0fab3b", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:56:56.799Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "02cd2c36-2bb9-413f-9758-946d6d16605b", "latest": "2026-04-29 16:56:52 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:57:09.059Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "4c569efa-6853-42f8-9f4d-5fd74c34239e", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T16:58:01.007Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "1ef9de58-0ce9-4a5e-87e1-5d4437a1ea71", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T16:58:07.684Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ecb843e8-86b6-45b4-82cf-c09e46925372", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T16:58:13.449Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "3605b3d7-4640-4394-8578-56e8030d7f52", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:19.147Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "da43c010-b2a6-49a9-8b75-d0030da22bb2", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:24.613Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "deb1b7a7-54d2-4b51-a86b-5f7ed2e57460", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:30.314Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "abdd8e96-9c7d-49f8-97d7-934633765e75", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:35.774Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "fbc19ad6-18a5-4789-bd06-2599520ff600", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:41.475Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e5d4a1bc-ea86-45a4-a463-c086582fd65e", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:42.224Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "3660853a-0459-422f-8bce-69e39f5751dd", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:47.965Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "84038480-d82a-4f9e-871b-ea8654c7faab", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:58:53.661Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e4ce600a-c75c-410c-a8db-b62d5675912f", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T16:59:04.289Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "28df4fc5-c86d-4109-afdb-91cf68ec78bd"} 2026-04-29T16:59:09.464Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6607e688-8dcc-45f0-b498-5257b76fe044"} 2026-04-29T16:59:13.931Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af2ea0e2-bde0-4aec-b4a0-54d79ec8523d", "user": "xtrabackup"} 2026-04-29T16:59:13.952Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af2ea0e2-bde0-4aec-b4a0-54d79ec8523d", "user": "xtrabackup"} 2026-04-29T16:59:13.972Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af2ea0e2-bde0-4aec-b4a0-54d79ec8523d", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T16:59:13.999Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af2ea0e2-bde0-4aec-b4a0-54d79ec8523d", "user": "xtrabackup"} 2026-04-29T16:59:14.015Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af2ea0e2-bde0-4aec-b4a0-54d79ec8523d", "user": "xtrabackup"} 2026-04-29T16:59:14.019Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af2ea0e2-bde0-4aec-b4a0-54d79ec8523d", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27"} 2026-04-29T16:59:14.025Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af2ea0e2-bde0-4aec-b4a0-54d79ec8523d", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777481934105871007", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiJkZjhjYmI5YmQ2NzM1MmMxMDIyOGY0ODcwNjkwZmU4MDIwYzE5MzNjNjkxMWY3Yjk0NmM0N2UyYjc2MDgwYjI3IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:56:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:58:54 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d", +  "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552",    "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-55dfc75967", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-55dfc75967", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:59:15.282Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "80cc8728-93cc-4a4b-a31e-4ca0967635d1"} 2026-04-29T16:59:45.902Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "2a0135e7-9d1a-45dc-a101-7f154a71f12d", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:59:48.213Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e021eec2-9c45-4ce7-a484-089d2d49d809", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:59:50.656Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7d9f77cf-6c6e-4ec5-970c-3eb8057f2c6a", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:59:53.124Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8c41311d-0987-4189-85ef-87fa5857c7cf", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:59:55.501Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "dd1d87b5-e33e-4f00-8c8b-15f7b56aec47", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:59:58.218Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "fc09aee7-ac04-45cc-a742-f198988db3fd", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:00:00.634Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "770b15f3-5201-49e7-8a8a-5a4fc94f62a1", "error": "exec binlog collector pod pitr-pitr-798584cd4f-jzrm4: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-798584cd4f-jzrm4\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:00:09.813Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c31a1004-fcb2-4cf3-9f97-591f3e236482", "latest": "2026-04-29 17:00:05 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:00:10.490Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "054df3b7-dee0-49b9-b975-74f46c83ac05", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:00:16.367Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "a617c886-ee8a-460d-915b-72772123d813", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:00:22.630Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "0c1454af-a67a-4a88-925b-dac759db95df", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:01:14.441Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "4f767f2b-ae9d-46f2-af12-2d96dd817727", "err": "failed to connect to pod pitr-pxc-0: dial tcp 10.158.161.61:33062: connect: connection refused"} 2026-04-29T17:01:19.757Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "726b5b4c-2175-4d3f-858a-def3e2febb9c", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:20.275Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "bfa40c1a-b4af-4cae-9b7f-4651b76e0e45", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:25.773Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "704eccd7-6d8d-45e9-a84a-8d6740af971d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:31.267Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "86dc53fe-dbe2-4191-b725-1cb6d6b44183", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:32.504Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "64298717-439d-4d83-8fb4-c9a802440b4a", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:37.984Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "998f72bf-cb68-4e60-a912-837be1affac8", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:43.694Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "473c3fbd-adb7-4c14-b65b-dfcdb205d1d4", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:49.425Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "66954e79-747e-470f-933a-ccd5c1b98d31", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:01:54.916Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "18a80455-2599-4535-b4cb-c38e07f92113", "primary name": "pitr-pxc-0.pitr-pxc.pitr-31942.svc.cluster.local"} 2026-04-29T17:02:04.347Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f409c5b5-936c-4867-8223-eb621963612a"} 2026-04-29T17:02:09.817Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f3e0f3de-8aa6-472c-aa53-cebef23f2b35"} 2026-04-29T17:02:15.061Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "54153833-50ea-4e14-ade5-938004102f0d"} 2026-04-29T17:02:20.779Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "51f2537e-ac6d-41ef-a617-1aa3dece3779"} 2026-04-29T17:02:26.585Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5bb38c38-6b21-4df6-91fd-ab6638d3519d"} 2026-04-29T17:02:32.676Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "3718fac2-8beb-48e4-900f-1a184a87d00a"} 2026-04-29T17:02:37.857Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "db2a40d7-3a01-4501-91ab-252a747dcbca"} 2026-04-29T17:02:43.563Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b9bb6fde-89b4-442c-9f6b-7ab87aa2fb87"} 2026-04-29T17:02:49.684Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f645b4ff-fa26-496c-9824-06eac9d8d730"} 2026-04-29T17:02:54.305Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ea4f0b4b-173d-4c63-a559-51b609041148", "latest": "2026-04-29 17:02:53 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:02:55.366Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ea4f0b4b-173d-4c63-a559-51b609041148"} 2026-04-29T17:03:01.064Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b8092ba4-4f9b-4a6f-9d6f-77c363b510bd"} 2026-04-29T17:03:06.672Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7ef74798-7b7a-4df2-b91c-3359210c36a9"} 2026-04-29T17:03:12.344Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "1fda6e96-49b8-4890-b269-1020b969ec69"} 2026-04-29T17:03:18.682Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e1f4eb87-c905-4fe1-bdbc-c6dc0092937b"} 2026-04-29T17:03:24.569Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "3404c1cb-923e-454b-9a49-7d689c76434e"} 2026-04-29T17:03:30.281Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "2885e663-b5da-4b95-96f4-119e42e1914a"} 2026-04-29T17:03:35.969Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "1b358a76-9eb2-41ae-bf04-8563029d619c"} 2026-04-29T17:03:41.788Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "51ad7f33-f16b-46e2-84f1-6e8adb5c9709"} 2026-04-29T17:03:47.775Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "160f8b8b-9b38-452d-ab5a-0c90ade09651"} 2026-04-29T17:03:53.439Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "07b75db6-df3d-4f56-8e9a-1466101fd701"} 2026-04-29T17:03:58.662Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b90823e5-7be1-4739-9ba9-31ecfe170aa5"} 2026-04-29T17:04:03.084Z INFO stopping cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7b0accc2-30a4-4ee6-aaf8-2080e0d840ca", "cluster": "pitr"} 2026-04-29T17:04:03.605Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "33111b4f-7676-48af-843b-3a82f0586a37", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777482116409423007", +  ResourceVersion: "", -  Generation: 6, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:59:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:01:56 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &3, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 6, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6cf587d4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6cf587d4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T17:04:03.650Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "33111b4f-7676-48af-843b-3a82f0586a37", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482087364479003", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:38 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:01:27 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-dff68c855", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-dff68c855", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T17:04:03.815Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "33111b4f-7676-48af-843b-3a82f0586a37", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482087364479003", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:45:38 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:01:27 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-dff68c855", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-dff68c855", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T17:04:04.648Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "56a0d6a1-4086-4081-80b2-0629ee0b3b5d", "error": "exec syncusers: failed to execute command in pod: pods \"pitr-proxysql-1\" not found / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: pods \"pitr-proxysql-1\" not found / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:993\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:834 2026-04-29T17:04:45.260Z INFO starting restore {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d6834c98-2ead-4566-af95-796b30fbf5ae", "cluster": "pitr", "backup": "on-pitr-minio"} 2026-04-29T17:04:45.331Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5d69a757-5a72-4cd3-b821-b7d3ca266b25", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:04:50.360Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3c8ba6f3-0bdb-48e7-b197-07f531d5092b", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:04:55.384Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d6a84c61-48f6-4905-b0df-21e6baeecf6e", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:05:00.404Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "52e12dbd-adc3-4c88-9d22-f5132db19ce2", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:05:05.424Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "53865297-fa4a-4fd1-a25c-3826b7c635f3", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:05:10.441Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "36469b83-a8eb-4902-bc45-bf8f0878ebbb", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:05:15.461Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c077d38f-44c7-49a4-bd29-b79e814c7b86", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:05:20.477Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d438b604-3ed8-4bde-9442-26a38a3e07a6", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:05:25.506Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d13b7f71-b71d-4a51-afda-6278ffb6224b", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:05:25.506Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d13b7f71-b71d-4a51-afda-6278ffb6224b", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:25.507Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d13b7f71-b71d-4a51-afda-6278ffb6224b", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:25.688Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "0f7141c7-6aa6-4d9b-b296-a9b1c60488b9", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777482283766655007", +  ResourceVersion: "", -  Generation: 7, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:04:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:04:43 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 7, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6cf587d4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6cf587d4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:05:25.741Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "0f7141c7-6aa6-4d9b-b296-a9b1c60488b9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482245835151003", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZDcyMTA3NjQ2NDQ2NWNjOTk5YmU3MWVlNTU4ZGE4NjkifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZDcyMTA3NjQ2NDQ2NWNjOTk5YmU3MWVlNTU4ZGE4NjkifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:04:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:04:05 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-dff68c855", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-dff68c855", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:05:25.819Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "0f7141c7-6aa6-4d9b-b296-a9b1c60488b9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482245835151003", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZDcyMTA3NjQ2NDQ2NWNjOTk5YmU3MWVlNTU4ZGE4NjkifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJlYzRiYzdjNTdiZWZhNWZmZDE3NTFkZGQ4YWMwOTU1MiIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZDcyMTA3NjQ2NDQ2NWNjOTk5YmU3MWVlNTU4ZGE4NjkifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:04:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:04:05 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-dff68c855", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-dff68c855", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:05:30.571Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f0ba1cb4-0d2d-48e2-b9d6-3d723c5f4cf5", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:05:30.571Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f0ba1cb4-0d2d-48e2-b9d6-3d723c5f4cf5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:30.571Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f0ba1cb4-0d2d-48e2-b9d6-3d723c5f4cf5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:30.571Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f0ba1cb4-0d2d-48e2-b9d6-3d723c5f4cf5", "cluster": "pitr"} 2026-04-29T17:05:35.587Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9d2b4afc-0e01-45b8-be7e-9428a6f0ee5b", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:05:35.587Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9d2b4afc-0e01-45b8-be7e-9428a6f0ee5b", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:35.587Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9d2b4afc-0e01-45b8-be7e-9428a6f0ee5b", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:35.587Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9d2b4afc-0e01-45b8-be7e-9428a6f0ee5b", "cluster": "pitr"} 2026-04-29T17:05:40.606Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8b26fa9b-519a-4d9e-9030-4971e56ee7e5", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:05:40.606Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8b26fa9b-519a-4d9e-9030-4971e56ee7e5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:40.607Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8b26fa9b-519a-4d9e-9030-4971e56ee7e5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:40.607Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8b26fa9b-519a-4d9e-9030-4971e56ee7e5", "cluster": "pitr"} 2026-04-29T17:05:45.623Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a37d91c1-f888-40d5-9ad0-65aa2d2022bd", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:05:45.623Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a37d91c1-f888-40d5-9ad0-65aa2d2022bd", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:45.624Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a37d91c1-f888-40d5-9ad0-65aa2d2022bd", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:45.624Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a37d91c1-f888-40d5-9ad0-65aa2d2022bd", "cluster": "pitr"} 2026-04-29T17:05:50.640Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "427a21aa-4cc9-4c39-878f-06fa0fad1e7d", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:05:50.640Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "427a21aa-4cc9-4c39-878f-06fa0fad1e7d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:50.640Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "427a21aa-4cc9-4c39-878f-06fa0fad1e7d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:50.640Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "427a21aa-4cc9-4c39-878f-06fa0fad1e7d", "cluster": "pitr"} 2026-04-29T17:05:55.661Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "09e07bf7-63d9-42d3-9f4b-62a74a75209e", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:05:55.661Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "09e07bf7-63d9-42d3-9f4b-62a74a75209e", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:55.661Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "09e07bf7-63d9-42d3-9f4b-62a74a75209e", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:05:55.661Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "09e07bf7-63d9-42d3-9f4b-62a74a75209e", "cluster": "pitr"} 2026-04-29T17:06:00.678Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "67c250bb-aacf-4d83-b3d7-4a9f166f1ffc", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:00.678Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "67c250bb-aacf-4d83-b3d7-4a9f166f1ffc", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:00.679Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "67c250bb-aacf-4d83-b3d7-4a9f166f1ffc", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:00.679Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "67c250bb-aacf-4d83-b3d7-4a9f166f1ffc", "cluster": "pitr"} 2026-04-29T17:06:05.697Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "65cc387c-bf1e-4576-a21e-cbc4698d7ac5", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:05.697Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "65cc387c-bf1e-4576-a21e-cbc4698d7ac5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:05.697Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "65cc387c-bf1e-4576-a21e-cbc4698d7ac5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:05.697Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "65cc387c-bf1e-4576-a21e-cbc4698d7ac5", "cluster": "pitr"} 2026-04-29T17:06:08.162Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "03e24e80-5fa2-49f8-a634-117d6d4ed43b"} 2026-04-29T17:06:08.238Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "03e24e80-5fa2-49f8-a634-117d6d4ed43b", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:06:10.714Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8e43b1a0-b6c2-4c57-8891-a7cfc2e65e52", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:10.714Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8e43b1a0-b6c2-4c57-8891-a7cfc2e65e52", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:10.715Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8e43b1a0-b6c2-4c57-8891-a7cfc2e65e52", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:10.715Z INFO point-in-time recovering {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8e43b1a0-b6c2-4c57-8891-a7cfc2e65e52", "cluster": "pitr"} 2026-04-29T17:06:10.783Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7fd3fe68-9a5a-40cc-9a62-925afdabb3b6", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:06:13.370Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "761b6ef2-26e7-4f37-8a4c-557fa2366e88"} 2026-04-29T17:06:14.237Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "761b6ef2-26e7-4f37-8a4c-557fa2366e88", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:06:15.803Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "122a8340-07b1-4ee6-96cf-157bd6145254", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:06:19.395Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e2df11ba-7a5f-41bf-9d0d-8912e7b2dba5", "user": "xtrabackup"} 2026-04-29T17:06:19.398Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e2df11ba-7a5f-41bf-9d0d-8912e7b2dba5"} 2026-04-29T17:06:19.485Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e2df11ba-7a5f-41bf-9d0d-8912e7b2dba5", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:06:20.844Z INFO starting cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e654ab20-b5f1-459e-b9f5-b51f031c9a7e", "cluster": "pitr"} 2026-04-29T17:06:21.061Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f4931a02-e5aa-4c2c-85af-048df5a2e879", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777482367079727007", +  ResourceVersion: "", -  Generation: 8, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:25 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:06:07 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &1, +  Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 8, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0, -  ReadyReplicas: 1, +  ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6cf587d4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6cf587d4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 1, +  AvailableReplicas: 0,    },   } 2026-04-29T17:06:21.156Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f4931a02-e5aa-4c2c-85af-048df5a2e879", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482325920239003", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:25 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:25 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-77bd7859b9", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-77bd7859b9", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:06:21.242Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f4931a02-e5aa-4c2c-85af-048df5a2e879", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482325920239003", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:25 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:25 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-77bd7859b9", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-77bd7859b9", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:06:24.051Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f4931a02-e5aa-4c2c-85af-048df5a2e879", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:06:25.936Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb21c009-e8c4-4c46-82a3-dfae491952b8", "cluster": "pitr"} 2026-04-29T17:06:27.202Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "9fc68b2d-d481-459a-aadb-5b143bdd1ac5", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:06:30.951Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8474169e-b2d6-4232-bf47-bd27d84fc162", "cluster": "pitr"} 2026-04-29T17:06:34.316Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6ac8bb17-db91-48c1-8662-fdecf5fffaff", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 10.158.160.72:6032: connect: connection refused"} 2026-04-29T17:06:35.970Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f3629971-1a48-40b8-9d63-6b5b67eb5ad2", "cluster": "pitr"} 2026-04-29T17:06:40.984Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a926a83c-34d8-440c-b891-425c1992f8f9", "cluster": "pitr"} 2026-04-29T17:06:45.999Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "32142615-babb-4961-9b4c-874c994d51e3", "cluster": "pitr"} 2026-04-29T17:06:51.016Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c627d408-0f4b-45f6-89ac-91624c1fd920", "cluster": "pitr"} 2026-04-29T17:06:56.030Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b32fa07e-fc1d-4cd9-aa66-bac0a455e033", "cluster": "pitr"} 2026-04-29T17:07:01.047Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "02d5525e-7cbd-4bf3-84e7-844aae4e9810", "cluster": "pitr"} 2026-04-29T17:07:06.066Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b104ae3d-6831-402e-a11b-13cfb88ee603", "cluster": "pitr"} 2026-04-29T17:07:11.081Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "020f5521-1421-4678-936a-db83f61e4289", "cluster": "pitr"} 2026-04-29T17:07:16.096Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "43ad6cd6-957a-4e7d-86b6-dac49ef984a2", "cluster": "pitr"} 2026-04-29T17:07:21.113Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a2fbbf64-ea18-4e56-9658-22c7063f7762", "cluster": "pitr"} 2026-04-29T17:07:26.127Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "bbe6da9f-b3be-4e45-93c3-fa1eece32836", "cluster": "pitr"} 2026-04-29T17:07:31.143Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "bd4e94b3-6668-45fe-ab98-d7fbd53bf325", "cluster": "pitr"} 2026-04-29T17:07:36.160Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e0b7980d-3e57-4fab-a206-3699bcb04402", "cluster": "pitr"} 2026-04-29T17:07:41.175Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "037d2df6-4b02-4c07-867c-34d69007fcff", "cluster": "pitr"} 2026-04-29T17:07:46.196Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ed41a185-5d5a-4360-b79f-2171f17f6355", "cluster": "pitr"} 2026-04-29T17:07:51.224Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c23f3d49-f8e1-4d45-8591-da4ca2acfecc", "cluster": "pitr"} 2026-04-29T17:07:56.242Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "76f0dda3-f95a-4e03-a68c-fe6bd6957430", "cluster": "pitr"} 2026-04-29T17:08:01.257Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fa2ed100-5c01-4d1b-b001-d05ae68e9653", "cluster": "pitr"} 2026-04-29T17:08:06.273Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0368e68c-1559-4227-87fa-1c69113e14d1", "cluster": "pitr"} 2026-04-29T17:08:11.292Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d48b3459-f366-4fa9-b5d3-00bbf0555093", "cluster": "pitr"} 2026-04-29T17:08:16.312Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7b4dd5ef-dfe0-4950-a49e-28e5836a36d7", "cluster": "pitr"} 2026-04-29T17:08:21.329Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "22a51956-ab70-4b15-9666-35358fa3d1e9", "cluster": "pitr"} 2026-04-29T17:08:26.346Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "25002402-c8cf-418a-9587-3d448b6700c0", "cluster": "pitr"} 2026-04-29T17:08:31.360Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fae46762-aa46-4e1b-8407-d6325bbed282", "cluster": "pitr"} 2026-04-29T17:08:36.379Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b8cf1230-59c1-4f4f-bfa9-5591255f6119", "cluster": "pitr"} 2026-04-29T17:08:41.400Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fbe0b6af-1136-4a65-836c-06c8f67beacf", "cluster": "pitr"} 2026-04-29T17:08:46.418Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ac0eb7b7-f075-4ed1-b067-b00ae4baceeb", "cluster": "pitr"} 2026-04-29T17:08:51.436Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "405a2111-976d-476e-8de9-64c2a54e4cad", "cluster": "pitr"} 2026-04-29T17:08:56.456Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1bd145d3-44a2-486e-989b-3cc2e0d9e45c", "cluster": "pitr"} 2026-04-29T17:09:01.469Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e9fa8e1e-2a51-4127-8c93-64b8147c6539", "cluster": "pitr"} 2026-04-29T17:09:04.247Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "9ba25186-c4ba-4920-88f7-0abf69e31baa"} 2026-04-29T17:09:08.122Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c052a96e-7a33-4a53-ac1f-7b09f9e2bd20", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T17:09:09.737Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c052a96e-7a33-4a53-ac1f-7b09f9e2bd20"} 2026-04-29T17:09:15.101Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f049a038-7530-437f-b183-a71597f8c92c"} 2026-04-29T17:09:19.575Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5463e4ac-2e43-4fbc-9caf-b959e60bd1e7", "latest": "2026-04-29 17:07:09 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:09:20.761Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5463e4ac-2e43-4fbc-9caf-b959e60bd1e7"} 2026-04-29T17:09:26.617Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "384d1d44-f978-4de6-a876-dbe9b3947e44"} 2026-04-29T17:09:32.662Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "9b84c224-516d-4876-bf39-6e4997d42f5b"} 2026-04-29T17:09:38.235Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "92e09b2f-2b08-4cfe-9bd5-eee02bb665b3"} 2026-04-29T17:09:44.013Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "313911ee-db08-4ee0-9768-a060a92bdb15"} 2026-04-29T17:09:49.559Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "037dc985-388b-42bd-9f51-e9d5d599e7db"} 2026-04-29T17:09:55.261Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "4bca9351-8269-463c-8074-147f6c30d053"} 2026-04-29T17:10:01.146Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "63715773-207b-4acf-a844-ac85515192b3"} 2026-04-29T17:10:06.755Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5fc8b1b8-6768-4fd7-91f6-e63c26fd8e5d"} 2026-04-29T17:10:12.726Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5d43b07f-167c-4607-b427-26e6be8ae429"} 2026-04-29T17:10:18.828Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c0e658cc-788c-4bf8-9d11-d021eb1633c5"} 2026-04-29T17:10:24.643Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e28eb874-612f-4a50-a73c-90f10a9a7bc1"} 2026-04-29T17:10:30.018Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "62a62f38-168e-402f-a6c1-7f9239a7f14b"} 2026-04-29T17:10:36.133Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "05d27e28-cae7-4702-ab64-e7b9978e1798"} 2026-04-29T17:10:42.136Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "edf8f7ca-29a2-4381-9c21-7138fa84130a"} 2026-04-29T17:10:47.736Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "43e6e0d4-5bb8-48ec-bf62-b96b89ceeaf3"} 2026-04-29T17:10:53.843Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "00f3b3b4-73de-4f4f-b404-6a7291b33834"} 2026-04-29T17:10:59.145Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e1fd5666-9e95-44e5-8854-d3792b1670a3"} 2026-04-29T17:11:05.117Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8cf1bf7e-c987-44e3-871e-a490ef707adc"} 2026-04-29T17:11:05.900Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "a7d6b8fb-6550-44b7-9f86-95b972020bc4", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:05.957Z INFO Created a new backup job {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "a7d6b8fb-6550-44b7-9f86-95b972020bc4", "namespace": "pitr-31942", "name": "xb-on-pitr-minio-2"} 2026-04-29T17:11:10.835Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "ad3cb01b-e600-4a27-abf1-190840e867cd"} 2026-04-29T17:11:10.958Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "261fea3c-cf3a-432c-bb33-a3982276d655", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:11.027Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "3323751e-78ff-41ea-8563-e94bed153b15", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:16.063Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "d9c7c384-fdb7-4f32-afe0-f547a85fdb92", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:16.248Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "13b7a377-e628-48b3-8914-7638ed928e80"} 2026-04-29T17:11:21.105Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "454c8e33-e0e3-4c5b-bbc0-984974682257", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:22.014Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "d246bd7e-b1ab-4c41-8410-73b6ae482030"} 2026-04-29T17:11:26.141Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "dfc1f5cf-0272-4db4-89e4-21a574487b79", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:27.905Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "9897e5ef-b105-4ddf-bb1d-baa6e28d3604"} 2026-04-29T17:11:31.221Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "a89501a6-bcf5-44c6-9efb-0e017383936a", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:33.813Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "0f7dcbb6-ec76-471b-be21-664a80b9b260"} 2026-04-29T17:11:36.257Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "4f690fd5-ceca-4e5c-a974-ea20a24f9103", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:39.650Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "2b54bbf9-9891-4523-a8e1-92b23e819a76"} 2026-04-29T17:11:41.294Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "f21fa0cb-ecc7-488e-9e25-7d395283bcf1", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:11:41.328Z INFO Backup succeeded {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "f21fa0cb-ecc7-488e-9e25-7d395283bcf1", "job": "xb-on-pitr-minio-2"} 2026-04-29T17:11:41.328Z DEBUG Removing binlog gap file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "f21fa0cb-ecc7-488e-9e25-7d395283bcf1", "job": "xb-on-pitr-minio-2", "pod": "pitr-pitr-798584cd4f-v66tp"} 2026-04-29T17:11:42.625Z DEBUG Removing binlog timeline file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "f21fa0cb-ecc7-488e-9e25-7d395283bcf1", "job": "xb-on-pitr-minio-2", "pod": "pitr-pitr-798584cd4f-v66tp"} 2026-04-29T17:11:42.907Z DEBUG Removing mysql-init secret {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "on-pitr-minio-2", "reconcileID": "f21fa0cb-ecc7-488e-9e25-7d395283bcf1", "job": "xb-on-pitr-minio-2", "secret": "pitr-mysql-init"} 2026-04-29T17:11:45.626Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e80f6350-1edd-4bf5-a697-eee0cce4006b"} 2026-04-29T17:11:51.517Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "2ce89170-36a1-4014-a45e-10bbdfd06a60"} 2026-04-29T17:11:56.947Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e390a515-53c6-4988-827f-ac3babff3980"} 2026-04-29T17:12:02.837Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c90c6b2c-b415-4776-8820-44648d81ddf0"} 2026-04-29T17:12:08.146Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "a7fcfbd1-9f4e-40d8-a877-445ec41678c5"} 2026-04-29T17:12:14.025Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "356fd1b2-e320-4723-b668-f12f3cdd45cc"} 2026-04-29T17:12:19.767Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "80628bad-2591-4b3f-aa1e-ae61c886d79b"} 2026-04-29T17:12:25.245Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5d7a87a8-294d-4807-81ec-a192f0d967ba"} 2026-04-29T17:12:31.156Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "26a2f595-9100-4317-b89e-daa20590f973"} 2026-04-29T17:12:36.812Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "278645f8-91cb-4e90-a286-8c3e7aa07f6d"} 2026-04-29T17:12:42.527Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "a5747226-bacd-4bb9-80f5-5e701f45c1c6"} 2026-04-29T17:12:48.537Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "18c7a2b0-415a-44bc-8d4c-d40e19317c27"} 2026-04-29T17:12:54.556Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "086ae38e-6162-427a-a8df-728ee2b271ab"} 2026-04-29T17:13:00.717Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "739d8c64-1b9f-4618-becf-8f8b06192339"} 2026-04-29T17:13:06.156Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "7bd2474e-feac-4ad7-8a17-441d919c6008"} 2026-04-29T17:13:11.850Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b3cbc6de-1a0b-416c-9080-9e3256caf19d"} 2026-04-29T17:13:17.923Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "9d319f2a-14c2-4014-a70e-5662f3db2723"} 2026-04-29T17:13:24.132Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "e7f7abd0-893d-445a-bf45-014e0a890eb9"} 2026-04-29T17:13:30.208Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "bdc92646-e3f9-4ec1-9fa2-a63f909b1a82"} 2026-04-29T17:13:35.618Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "fd5913f8-1406-45ba-a0b2-4a1299459a15"} 2026-04-29T17:13:41.242Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "590e41fb-79ae-4b12-a7eb-4b995c5c3b4c"} 2026-04-29T17:13:43.067Z INFO stopping cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "370c0445-dec1-4a28-b6b6-49f41b8286fb", "cluster": "pitr"} 2026-04-29T17:13:43.192Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c9c555c0-f7bb-4335-82ff-580815772869"} 2026-04-29T17:13:43.193Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c9c555c0-f7bb-4335-82ff-580815772869", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777482541398687007", +  ResourceVersion: "", -  Generation: 9, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:06:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:09:01 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &3, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 9, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6cf587d4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6cf587d4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T17:13:43.252Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c9c555c0-f7bb-4335-82ff-580815772869", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482416922895003", +  ResourceVersion: "", -  Generation: 6, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:06:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:06:56 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 6, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-77bd7859b9", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-77bd7859b9", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T17:13:43.386Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c9c555c0-f7bb-4335-82ff-580815772869", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482416922895003", +  ResourceVersion: "", -  Generation: 6, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:06:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:06:56 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 6, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-77bd7859b9", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-77bd7859b9", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T17:14:24.257Z INFO starting restore {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "3daa5ae0-a326-43c0-afc1-9a25c3d7907b", "cluster": "pitr", "backup": "on-pitr-minio-2"} 2026-04-29T17:14:24.331Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "752d5f57-7b06-4ffa-8423-29fd1598b6d9", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:14:29.349Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "92d18571-df4d-40a7-8219-53795fe6d5b2", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:14:34.364Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "c9ef9652-b659-411b-a43e-49942ed3afdb", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:14:39.391Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "eb173737-ed0e-4806-9bc6-7d3f81a639c0", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:14:44.404Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "8138e636-841f-488c-ac9e-89892ddbead2", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:14:49.424Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "0204575b-4135-414b-b727-099d6fdee45e", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:14:49.424Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "0204575b-4135-414b-b727-099d6fdee45e", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:14:49.425Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "0204575b-4135-414b-b727-099d6fdee45e", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:14:49.589Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "06af3f76-23f0-4075-976b-c1caef337724", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777482862926671007", +  ResourceVersion: "", -  Generation: 10, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:43 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:14:22 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 10, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6cf587d4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6cf587d4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:14:49.638Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "06af3f76-23f0-4075-976b-c1caef337724", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777482862926671007", +  ResourceVersion: "", -  Generation: 10, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:43 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:14:22 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 10, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6cf587d4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6cf587d4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:14:54.488Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "969354fd-a6fe-4ef1-81ca-510b0ff7074a", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:14:54.488Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "969354fd-a6fe-4ef1-81ca-510b0ff7074a", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:14:54.488Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "969354fd-a6fe-4ef1-81ca-510b0ff7074a", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:14:54.488Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "969354fd-a6fe-4ef1-81ca-510b0ff7074a", "cluster": "pitr"} 2026-04-29T17:14:59.516Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "36b9c574-3faf-47ee-8816-2c66d50c4a7f", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:14:59.516Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "36b9c574-3faf-47ee-8816-2c66d50c4a7f", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:14:59.516Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "36b9c574-3faf-47ee-8816-2c66d50c4a7f", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:14:59.516Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "36b9c574-3faf-47ee-8816-2c66d50c4a7f", "cluster": "pitr"} 2026-04-29T17:15:04.534Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "aa442786-9fba-4a5a-9f53-67ed53f10218", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:04.534Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "aa442786-9fba-4a5a-9f53-67ed53f10218", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:04.535Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "aa442786-9fba-4a5a-9f53-67ed53f10218", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:04.535Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "aa442786-9fba-4a5a-9f53-67ed53f10218", "cluster": "pitr"} 2026-04-29T17:15:09.550Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "babebf27-e5b7-41a4-b8cd-d7f355aab33b", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:09.550Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "babebf27-e5b7-41a4-b8cd-d7f355aab33b", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:09.550Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "babebf27-e5b7-41a4-b8cd-d7f355aab33b", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:09.550Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "babebf27-e5b7-41a4-b8cd-d7f355aab33b", "cluster": "pitr"} 2026-04-29T17:15:14.565Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "cb1b1780-6f11-4478-a4b8-cc8d78065a15", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:14.565Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "cb1b1780-6f11-4478-a4b8-cc8d78065a15", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:14.566Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "cb1b1780-6f11-4478-a4b8-cc8d78065a15", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:14.566Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "cb1b1780-6f11-4478-a4b8-cc8d78065a15", "cluster": "pitr"} 2026-04-29T17:15:19.583Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "bd98119e-3a50-403b-813f-94ef54ab02c2", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:19.583Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "bd98119e-3a50-403b-813f-94ef54ab02c2", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:19.583Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "bd98119e-3a50-403b-813f-94ef54ab02c2", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:19.583Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "bd98119e-3a50-403b-813f-94ef54ab02c2", "cluster": "pitr"} 2026-04-29T17:15:24.597Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "8d620ea5-1052-4d00-97cb-9ceeee3a8f75", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:24.597Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "8d620ea5-1052-4d00-97cb-9ceeee3a8f75", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:24.598Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "8d620ea5-1052-4d00-97cb-9ceeee3a8f75", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:24.598Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "8d620ea5-1052-4d00-97cb-9ceeee3a8f75", "cluster": "pitr"} 2026-04-29T17:15:29.614Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "fc94f472-af45-4028-9de2-3818c43e3e11", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:29.614Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "fc94f472-af45-4028-9de2-3818c43e3e11", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:29.614Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "fc94f472-af45-4028-9de2-3818c43e3e11", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:29.614Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "fc94f472-af45-4028-9de2-3818c43e3e11", "cluster": "pitr"} 2026-04-29T17:15:34.636Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "217ecddc-d1cc-4f36-90e2-ebcd45ad5fc9", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:34.636Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "217ecddc-d1cc-4f36-90e2-ebcd45ad5fc9", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:34.636Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "217ecddc-d1cc-4f36-90e2-ebcd45ad5fc9", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:34.636Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "217ecddc-d1cc-4f36-90e2-ebcd45ad5fc9", "cluster": "pitr"} 2026-04-29T17:15:39.653Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "d4ac1caa-0c23-4d61-9ed4-063c2f0f5bfc", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:39.653Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "d4ac1caa-0c23-4d61-9ed4-063c2f0f5bfc", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:39.653Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "d4ac1caa-0c23-4d61-9ed4-063c2f0f5bfc", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:39.653Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "d4ac1caa-0c23-4d61-9ed4-063c2f0f5bfc", "cluster": "pitr"} 2026-04-29T17:15:41.420Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "caae1511-7941-42bb-b530-0ab37f9ad044"} 2026-04-29T17:15:41.925Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "caae1511-7941-42bb-b530-0ab37f9ad044", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:15:44.674Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "9a3117c7-d4b5-4882-8147-c59315fea954", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:15:44.674Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "9a3117c7-d4b5-4882-8147-c59315fea954", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:44.674Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "9a3117c7-d4b5-4882-8147-c59315fea954", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:15:44.674Z INFO point-in-time recovering {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "9a3117c7-d4b5-4882-8147-c59315fea954", "cluster": "pitr"} 2026-04-29T17:15:44.732Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "06975412-f50b-44da-ba99-8ab5fa6d0832", "job": "pitr-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:15:47.200Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6bca80e4-50a0-4c19-bc73-69349c174acb"} 2026-04-29T17:15:47.259Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6bca80e4-50a0-4c19-bc73-69349c174acb", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:15:49.756Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "bb8b89ad-f12b-4747-a85a-eede96565f42", "job": "pitr-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:15:52.408Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5bf23f84-aacc-4999-a08a-f172a2263b99"} 2026-04-29T17:15:52.878Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5bf23f84-aacc-4999-a08a-f172a2263b99", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:15:54.773Z INFO starting cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "6c95499f-fdde-4b3a-bdac-1ff9b9a6a41b", "cluster": "pitr"} 2026-04-29T17:15:55.024Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8b285276-2f78-4b91-b696-312e899e3df0", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "fea1306e-79ab-4d3c-abd9-1df0b55c171d", +  UID: "", -  ResourceVersion: "1777482938054015007", +  ResourceVersion: "", -  Generation: 11, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:14:49 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:15:38 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &1, +  Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 11, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0, -  ReadyReplicas: 1, +  ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-6cf587d4c", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-6cf587d4c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 1, +  AvailableReplicas: 0,    },   } 2026-04-29T17:15:55.075Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8b285276-2f78-4b91-b696-312e899e3df0", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482825839135003", +  ResourceVersion: "", -  Generation: 7, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:43 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:45 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 7, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-77bd7859b9", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-77bd7859b9", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:15:55.137Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8b285276-2f78-4b91-b696-312e899e3df0", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-31942",    SelfLink: "", -  UID: "5cade122-4da7-4f4c-934f-244e4a5a361f", +  UID: "", -  ResourceVersion: "1777482825839135003", +  ResourceVersion: "", -  Generation: 7, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:45:30 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "d27c8b4c-11c3-4b0d-81f6-454cfb318f17", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:43 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:45 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "ec4bc7c57befa5ffd1751ddd8ac09552", "percona.com/ssl-internal-hash": "d721076464465cc999be71ee558da869"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 7, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-77bd7859b9", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-77bd7859b9", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:15:57.766Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8b285276-2f78-4b91-b696-312e899e3df0", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:15:59.882Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "1856eae2-3102-4094-b641-3d7d8b6a7763", "cluster": "pitr"} 2026-04-29T17:16:01.292Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "00905edc-456d-469d-b731-e10dc8816491", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-31942 on 34.118.224.10:53: no such host"} 2026-04-29T17:16:04.912Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "b9a22f14-a473-49ac-9541-9cfff41a362a", "cluster": "pitr"} 2026-04-29T17:16:09.928Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "fa5c0c4e-8bf3-4103-9c4e-6cc371cfc23f", "cluster": "pitr"} 2026-04-29T17:16:14.946Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "c93da722-e3b1-4883-9080-e678facad5b8", "cluster": "pitr"} 2026-04-29T17:16:19.964Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "6c7138d4-91b7-4208-9cf2-0bb88562e2bb", "cluster": "pitr"} 2026-04-29T17:16:24.987Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "4d826922-6b94-43d8-acf7-3624cb7ef142", "cluster": "pitr"} 2026-04-29T17:16:30.003Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "e4be3ab5-cbad-42b6-8a2b-01aa39522cbb", "cluster": "pitr"} 2026-04-29T17:16:35.023Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "c1224717-3d82-47f2-93c0-bc1227511623", "cluster": "pitr"} 2026-04-29T17:16:40.042Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "761f734e-06b2-478c-923c-cccfc6474f4b", "cluster": "pitr"} 2026-04-29T17:16:45.058Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "f7cdb594-aec5-4d2b-8fc0-1440e9031dd0", "cluster": "pitr"} 2026-04-29T17:16:50.073Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "3ebdddd7-fd19-4bb2-864f-a1bc98aa10cb", "cluster": "pitr"} 2026-04-29T17:16:55.094Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "24cb2aed-358f-45cc-b60c-49c6229d839e", "cluster": "pitr"} 2026-04-29T17:17:00.110Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "d8b7e44b-2c48-42fb-8100-7e1a1baac5a5", "cluster": "pitr"} 2026-04-29T17:17:05.131Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "93fee8a5-460a-4765-bf29-20337d55f649", "cluster": "pitr"} 2026-04-29T17:17:10.149Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "1d1c0f60-7331-4531-8bf4-6099e697752f", "cluster": "pitr"} 2026-04-29T17:17:15.165Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "71cbe6d3-d0de-43ec-9767-57b986dd5bb4", "cluster": "pitr"} 2026-04-29T17:17:20.181Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "95172469-b006-4864-9849-8d786859ab94", "cluster": "pitr"} 2026-04-29T17:17:25.200Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "5cfb2cbb-7b62-44d1-ba68-33ca0070817e", "cluster": "pitr"} 2026-04-29T17:17:30.216Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "27012438-c954-4766-8c4b-bbbfc607376a", "cluster": "pitr"} 2026-04-29T17:17:35.235Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "b9d30659-dcfe-42c2-bc98-af97e350e257", "cluster": "pitr"} 2026-04-29T17:17:40.251Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "60246561-3aca-4548-94a6-ffbc8bc87499", "cluster": "pitr"} 2026-04-29T17:17:45.271Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "07c9432f-5587-42f8-a1c3-e42d495d0fdc", "cluster": "pitr"} 2026-04-29T17:17:50.288Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "cbe6c392-c698-4e19-a615-1b9a0554b8d8", "cluster": "pitr"} 2026-04-29T17:17:55.307Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "c60ff3e8-cc45-471a-a0da-61b68a74d8d9", "cluster": "pitr"} 2026-04-29T17:18:00.321Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "2e27930b-1983-4d8e-8028-8df6974282a8", "cluster": "pitr"} 2026-04-29T17:18:05.336Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "9569c886-2a02-4f33-b909-d3fa894de440", "cluster": "pitr"} 2026-04-29T17:18:10.353Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "8f79df53-42c9-4b73-9dd4-b1d0322c9a6f", "cluster": "pitr"} 2026-04-29T17:18:15.381Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "98d20d1e-db85-4f6e-a0c6-c25d47543a74", "cluster": "pitr"} 2026-04-29T17:18:20.396Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "b8f6ac42-20b9-4f0d-9570-19fa9c14f228", "cluster": "pitr"} 2026-04-29T17:18:25.412Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "3363443f-c4ac-414a-8373-36c41d384432", "cluster": "pitr"} 2026-04-29T17:18:30.427Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "restore-on-pitr-minio-time", "reconcileID": "b608a66b-f31b-4029-87ba-3fa061b6cd11", "cluster": "pitr"} 2026-04-29T17:18:36.057Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "5e05ee38-ed99-4314-9287-d6cde8647197"} 2026-04-29T17:18:39.526Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "850f624d-7dbe-4469-b388-9a8923a0c1c9", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T17:18:41.443Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "850f624d-7dbe-4469-b388-9a8923a0c1c9"} 2026-04-29T17:18:46.459Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "af6c0dba-f062-46a5-9934-86b7e4b2a923"} 2026-04-29T17:18:51.940Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6b9915bb-59e6-42e7-96e9-66936adbb074"} 2026-04-29T17:18:57.965Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6ae9b5e6-c4b8-4859-81bf-021347db0370"} 2026-04-29T17:19:03.761Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "0b7c86bb-b3f4-4c90-aa24-44986d0242e5"} 2026-04-29T17:19:09.872Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "33ac27ef-5754-4f34-a01c-0e38e7680809"} 2026-04-29T17:19:15.551Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f281edbd-26e4-41b6-b315-e50ad2db6f68"} 2026-04-29T17:19:21.246Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "db444102-5673-4a2b-905d-9891434dbe2f"} 2026-04-29T17:19:27.143Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "d4ac9e1d-66ce-4fe5-a964-304767a42cd9"} 2026-04-29T17:19:33.052Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f5ec9307-272f-4e76-a92e-882ee62eed02"} 2026-04-29T17:19:39.179Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "17f315df-2037-4fbb-8749-3bf7fbcc16cc"} 2026-04-29T17:19:44.944Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8c107908-9fab-4278-83ca-57c22465fe54"} 2026-04-29T17:19:50.360Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "21a8e0d1-0920-4934-933f-481557dfd40a"} 2026-04-29T17:19:56.537Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "926678ad-e213-4259-8e1a-38f126948c08"} 2026-04-29T17:20:02.575Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "8a4a10ef-8fe4-48c8-bb14-37912ed9db58"} 2026-04-29T17:20:08.059Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "12a8aaa8-d692-4a11-8372-94949b0517d2"} 2026-04-29T17:20:14.156Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "c4f65e63-9214-4129-b11c-62d39b858c1e"} 2026-04-29T17:20:19.871Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "d51ac10b-6bdb-42ee-9a78-fa87f1ce269f"} 2026-04-29T17:20:25.549Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b552547a-702b-4eae-b330-f1565e3be402"} 2026-04-29T17:20:31.707Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "24541def-e1de-4385-8da8-378a305b066d"} 2026-04-29T17:20:37.548Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "fe61b551-fcf8-4cdf-bf47-c5a35d79629c"} 2026-04-29T17:20:43.059Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "2d534e62-c786-4d8b-84f7-38ddc18e1bc9"} 2026-04-29T17:20:48.943Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b65be728-7a6c-45e6-9441-1d5a5a6a5e7e"} 2026-04-29T17:20:54.452Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f46540ed-578d-4907-9fea-d97b3214c930"} 2026-04-29T17:21:00.647Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "95dd9327-826e-42ea-8a7d-ed48eeda86ee"} 2026-04-29T17:21:06.259Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "55216fc2-9aa5-4d22-aa1d-987f335a7d27"} 2026-04-29T17:21:11.959Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f54af1f9-14ee-4dd8-9b60-032faff8413a"} 2026-04-29T17:21:17.727Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "d6062fa7-8837-4986-b508-ba39189ed1e4"} 2026-04-29T17:21:23.267Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "677c2868-3e57-4897-bf0a-42ab78b1ca6b"} 2026-04-29T17:21:29.257Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "aacf666c-58f2-4a19-84ca-7dea0a053ed7"} 2026-04-29T17:21:34.853Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "724d6531-2353-47fa-9fac-c446b19dffc4"} 2026-04-29T17:21:40.856Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "53434faa-a3e0-4ad9-8d13-4f467c5c4384"} 2026-04-29T17:21:46.444Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "d12e8808-4fe8-4fe9-8ae9-0f64b36f8604"} 2026-04-29T17:21:52.348Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "b83946ce-1612-49e5-bdbb-51822e07bd0d"} 2026-04-29T17:21:58.172Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "6d5b1bd6-5209-4a60-9d81-5541647bafc7"} 2026-04-29T17:22:03.653Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "3278fd1f-c168-485c-b0b4-ac0da76754b3"} 2026-04-29T17:22:09.970Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "731f4da4-49a7-482d-811f-a9de5d5b3039"} 2026-04-29T17:22:15.671Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-31942"}, "namespace": "pitr-31942", "name": "pitr", "reconcileID": "f948bd1c-da55-4f1b-a6a6-2de32e55161b"} max retry count 120 reached. something went wrong with operator or kubernetes cluster