Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/logs/pitr-8-0.log Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 + GTID_PATTERN='[A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + main + create_infra pitr-18261 + local ns=pitr-18261 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n pitr-10673 pitr --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/pitr patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.3cfkeRpvoi ++ mktemp + local LAST_ERR=/tmp/tmp.MG0l237iND + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3cfkeRpvoi perconaxtradbcluster.pxc.percona.com "pitr" deleted from pitr-10673 namespace + cat /tmp/tmp.MG0l237iND + rm /tmp/tmp.3cfkeRpvoi /tmp/tmp.MG0l237iND + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.YsYCMHScVL ++ mktemp + local LAST_ERR=/tmp/tmp.4APkWlnhWc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YsYCMHScVL perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio" deleted from pitr-10673 namespace perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio-2" deleted from pitr-10673 namespace + cat /tmp/tmp.4APkWlnhWc + rm /tmp/tmp.YsYCMHScVL /tmp/tmp.4APkWlnhWc + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.uKQSaAjTuL ++ mktemp + local LAST_ERR=/tmp/tmp.5JzGVaESlI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uKQSaAjTuL perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio-time" deleted from pitr-10673 namespace + cat /tmp/tmp.5JzGVaESlI + rm /tmp/tmp.uKQSaAjTuL /tmp/tmp.5JzGVaESlI + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl api-resources ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.Muc2guoIKK + kubectl_bin get ns + xargs kubectl delete ns + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.VcR24RebGZ + local LAST_ERR=/tmp/tmp.CWKxaSIhqK + local exit_status=0 ++ mktemp + local LAST_ERR=/tmp/tmp.fRL3hodl0m + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VcR24RebGZ + cat /tmp/tmp.fRL3hodl0m + rm /tmp/tmp.VcR24RebGZ /tmp/tmp.fRL3hodl0m + return 0 namespace "cert-manager" deleted namespace "pitr-10673" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Muc2guoIKK namespace "pxc-operator" deleted + cat /tmp/tmp.CWKxaSIhqK + rm /tmp/tmp.Muc2guoIKK /tmp/tmp.CWKxaSIhqK + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.SjiVWkYjVa ++ mktemp + local LAST_ERR=/tmp/tmp.CHHlRX7l1z + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SjiVWkYjVa namespace/pxc-operator created + cat /tmp/tmp.CHHlRX7l1z + rm /tmp/tmp.SjiVWkYjVa /tmp/tmp.CHHlRX7l1z + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.2uG4vONMrN +++ mktemp ++ local LAST_ERR=/tmp/tmp.FLrcXKu3M8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2uG4vONMrN ++ cat /tmp/tmp.FLrcXKu3M8 ++ rm /tmp/tmp.2uG4vONMrN /tmp/tmp.FLrcXKu3M8 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster4 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.CC84dcPlpG ++ mktemp + local LAST_ERR=/tmp/tmp.moanX6ZaBW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster4 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CC84dcPlpG Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster4" modified. + cat /tmp/tmp.moanX6ZaBW + rm /tmp/tmp.CC84dcPlpG /tmp/tmp.moanX6ZaBW + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.JxtUuEZtSc ++ mktemp + local LAST_ERR=/tmp/tmp.UVFk9jCj9z + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JxtUuEZtSc customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.UVFk9jCj9z + rm /tmp/tmp.JxtUuEZtSc /tmp/tmp.UVFk9jCj9z + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + kubectl_bin apply -f - + sed -e 's^namespace: .*^namespace: pxc-operator^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-rbac.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.CfRybLiXnD ++ mktemp + local LAST_ERR=/tmp/tmp.UBTDSa1v0M + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CfRybLiXnD clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.UBTDSa1v0M + rm /tmp/tmp.CfRybLiXnD /tmp/tmp.UBTDSa1v0M + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = ""' - + kubectl_bin apply -f - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.rlquczDiPt ++ mktemp + local LAST_ERR=/tmp/tmp.8x4lnPQRGC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rlquczDiPt deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.8x4lnPQRGC + rm /tmp/tmp.rlquczDiPt /tmp/tmp.8x4lnPQRGC + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.zsu4TEzkBE ++ mktemp + local LAST_ERR=/tmp/tmp.xrf2fgUZTr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zsu4TEzkBE pod/percona-xtradb-cluster-operator-7cd7bb76f4-4zmbw condition met + cat /tmp/tmp.xrf2fgUZTr + rm /tmp/tmp.zsu4TEzkBE /tmp/tmp.xrf2fgUZTr + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ head -1 ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qAGDQwIli5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FcsWbFq0I4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qAGDQwIli5 ++ cat /tmp/tmp.FcsWbFq0I4 ++ rm /tmp/tmp.qAGDQwIli5 /tmp/tmp.FcsWbFq0I4 ++ return 0 + wait_pod percona-xtradb-cluster-operator-7cd7bb76f4-4zmbw 480 pxc-operator + local pod=percona-xtradb-cluster-operator-7cd7bb76f4-4zmbw + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-7cd7bb76f4-4zmbw ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-7cd7bb76f4-4zmbw condition met waiting for pod/percona-xtradb-cluster-operator-7cd7bb76f4-4zmbw to become Ready.Ok + sleep 3 + create_namespace pitr-18261 + local namespace=pitr-18261 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pitr-18261' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pitr-18261 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pitr-18261 + xargs kubectl delete ns + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + kubectl_bin get ns + local LAST_OUT=/tmp/tmp.nuqpa0L7RN ++ mktemp + local LAST_OUT=/tmp/tmp.b1qrwMNtRj ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.vWSg4ZpMe7 + local exit_status=0 + local LAST_ERR=/tmp/tmp.53EBnso7fW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-18261 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.b1qrwMNtRj + cat /tmp/tmp.vWSg4ZpMe7 + rm /tmp/tmp.b1qrwMNtRj /tmp/tmp.vWSg4ZpMe7 + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-18261 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-18261 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.nuqpa0L7RN + cat /tmp/tmp.53EBnso7fW Error from server (NotFound): namespaces "pitr-18261" not found + rm /tmp/tmp.nuqpa0L7RN /tmp/tmp.53EBnso7fW + return 1 + : + wait_for_delete namespace/pitr-18261 + local res=namespace/pitr-18261 + echo -n 'waiting for namespace/pitr-18261 to be deleted' waiting for namespace/pitr-18261 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pitr-18261" not found + desc 'create namespace pitr-18261' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pitr-18261 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pitr-18261 ++ mktemp + local LAST_OUT=/tmp/tmp.6L8A8ktHFz ++ mktemp + local LAST_ERR=/tmp/tmp.G20oPdWL1R + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pitr-18261 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6L8A8ktHFz namespace/pitr-18261 created + cat /tmp/tmp.G20oPdWL1R + rm /tmp/tmp.6L8A8ktHFz /tmp/tmp.G20oPdWL1R + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.0V3JbbMnQ1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zliECnamtr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0V3JbbMnQ1 ++ cat /tmp/tmp.zliECnamtr ++ rm /tmp/tmp.0V3JbbMnQ1 /tmp/tmp.zliECnamtr ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster4 --namespace=pitr-18261 ++ mktemp + local LAST_OUT=/tmp/tmp.Cwa4nj4EuX ++ mktemp + local LAST_ERR=/tmp/tmp.87RW3VxmWP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster4 --namespace=pitr-18261 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Cwa4nj4EuX Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster4" modified. + cat /tmp/tmp.87RW3VxmWP + rm /tmp/tmp.Cwa4nj4EuX /tmp/tmp.87RW3VxmWP + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.rMKvEcigqE ++ mktemp + local LAST_ERR=/tmp/tmp.piEkQ7IHAe + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rMKvEcigqE secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.piEkQ7IHAe + rm /tmp/tmp.rMKvEcigqE /tmp/tmp.piEkQ7IHAe + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.8lVFWO1quy ++ mktemp + local LAST_ERR=/tmp/tmp.0MIhBias17 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.8lVFWO1quy namespace/cert-manager created + cat /tmp/tmp.0MIhBias17 + rm /tmp/tmp.8lVFWO1quy /tmp/tmp.0MIhBias17 + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.JhwKF8xgkH ++ mktemp + local LAST_ERR=/tmp/tmp.SiZwt4Egvp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JhwKF8xgkH namespace/cert-manager labeled + cat /tmp/tmp.SiZwt4Egvp + rm /tmp/tmp.JhwKF8xgkH /tmp/tmp.SiZwt4Egvp + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.gh7JzscVwf ++ mktemp + local LAST_ERR=/tmp/tmp.j55ymBWODo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.gh7JzscVwf namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.j55ymBWODo Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.gh7JzscVwf /tmp/tmp.j55ymBWODo + return 0 + '[' '' == 4.10 ']' + sleep 70 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/issuer.yml ++ mktemp + local LAST_OUT=/tmp/tmp.eBSD0mdN5r ++ mktemp + local LAST_ERR=/tmp/tmp.qVlakD4H9T + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/issuer.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eBSD0mdN5r issuer.cert-manager.io/selfsigning-issuer created + cat /tmp/tmp.qVlakD4H9T + rm /tmp/tmp.eBSD0mdN5r /tmp/tmp.qVlakD4H9T + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/cert.yml ++ mktemp + local LAST_OUT=/tmp/tmp.nSKlojwYSC ++ mktemp + local LAST_ERR=/tmp/tmp.z1S792r4sA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/cert.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nSKlojwYSC certificate.cert-manager.io/tls-minio created + cat /tmp/tmp.z1S792r4sA Warning: spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. + rm /tmp/tmp.nSKlojwYSC /tmp/tmp.z1S792r4sA + return 0 + sleep_with_log 25 + local d=25 + log 'sleeping for 25 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:32:53+0000]' sleeping for 25 seconds [2026-04-29T16:32:53+0000] sleeping for 25 seconds + sleep 25 + start_minio tls-minio + deploy_helm pitr-18261 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=tls-minio + local endpoint=http://minio-service:9000 + minio_args=(--version $MINIO_VER --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set "users[0].accessKey=some-access-key" --set "users[0].secretKey=some-secret-key" --set "users[0].policy=consoleAdmin" --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G) + local minio_args + [[ -n tls-minio ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Wed Apr 29 16:33:25 2026 NAMESPACE: pitr-18261 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.pitr-18261.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace pitr-18261 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace pitr-18261 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace pitr-18261 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace pitr-18261 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.la2NIujsP2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ApaHXMcjnf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.la2NIujsP2 ++ cat /tmp/tmp.ApaHXMcjnf ++ rm /tmp/tmp.la2NIujsP2 /tmp/tmp.ApaHXMcjnf ++ return 0 + MINIO_POD=minio-service-757b74c96c-2vzxj + wait_pod minio-service-757b74c96c-2vzxj + local pod=minio-service-757b74c96c-2vzxj + local max_retry=480 + local ns= ++ echo minio-service-757b74c96c-2vzxj ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/minio-service-757b74c96c-2vzxj condition met waiting for pod/minio-service-757b74c96c-2vzxj to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.N2jIWugiYR ++ mktemp + local LAST_ERR=/tmp/tmp.3qfgaZCQ5J + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.N2jIWugiYR /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from pitr-18261 namespace + cat /tmp/tmp.3qfgaZCQ5J + rm /tmp/tmp.N2jIWugiYR /tmp/tmp.3qfgaZCQ5J + return 0 + cluster=pitr + spinup_pxc pitr /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local cluster=pitr + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.KM9esxnuGi ++ mktemp + local LAST_ERR=/tmp/tmp.uu6G6jVulv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KM9esxnuGi secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.uu6G6jVulv + rm /tmp/tmp.KM9esxnuGi /tmp/tmp.uu6G6jVulv + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-18261~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local LAST_OUT=/tmp/tmp.vYG2Yi4OMs ++ mktemp + local LAST_ERR=/tmp/tmp.4pW2i33oxi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vYG2Yi4OMs deployment.apps/pxc-client created + cat /tmp/tmp.4pW2i33oxi + rm /tmp/tmp.vYG2Yi4OMs /tmp/tmp.4pW2i33oxi + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + local pvc_name= + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-18261~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/pitr.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ mktemp + local LAST_OUT=/tmp/tmp.6WUWyVVCnl ++ mktemp + local LAST_ERR=/tmp/tmp.3xo6GpT0T4 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6WUWyVVCnl perconaxtradbcluster.pxc.percona.com/pitr created + cat /tmp/tmp.3xo6GpT0T4 + rm /tmp/tmp.6WUWyVVCnl /tmp/tmp.3xo6GpT0T4 + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.aYA60Auv8u ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Rnlhmamkxj +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.aYA60Auv8u +++ cat /tmp/tmp.Rnlhmamkxj +++ rm /tmp/tmp.aYA60Auv8u /tmp/tmp.Rnlhmamkxj +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rE8rXFGhVg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uM1cuap3Zz +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.rE8rXFGhVg +++ cat /tmp/tmp.uM1cuap3Zz +++ rm /tmp/tmp.rE8rXFGhVg /tmp/tmp.uM1cuap3Zz +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-18261 ++ mktemp + local LAST_OUT=/tmp/tmp.aQy0v9a0ff ++ mktemp + local LAST_ERR=/tmp/tmp.iQ88OlJhHV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-18261 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-18261 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-18261 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.aQy0v9a0ff + cat /tmp/tmp.iQ88OlJhHV error: no matching resources found + rm /tmp/tmp.aQy0v9a0ff /tmp/tmp.iQ88OlJhHV + return 1 + true + wait_for_running pitr-proxysql 1 + local name=pitr-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-proxysql-0 ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/pitr-proxysql-0 condition met waiting for pod/pitr-proxysql-0 to become Ready.Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-0 + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-1 + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc pitr -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ base64 --decode ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.71wV1I2duF +++ mktemp ++ local LAST_ERR=/tmp/tmp.wgIW91UPNh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.71wV1I2duF ++ cat /tmp/tmp.wgIW91UPNh ++ rm /tmp/tmp.71wV1I2duF /tmp/tmp.wgIW91UPNh ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bKpHm6rxDy +++ mktemp ++ local LAST_ERR=/tmp/tmp.xiuQAUU97W ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bKpHm6rxDy ++ cat /tmp/tmp.xiuQAUU97W ++ rm /tmp/tmp.bKpHm6rxDy /tmp/tmp.xiuQAUU97W ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gp3YPhpTUU +++ mktemp ++ local LAST_ERR=/tmp/tmp.X2FNb8sBd7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gp3YPhpTUU ++ cat /tmp/tmp.X2FNb8sBd7 ++ rm /tmp/tmp.gp3YPhpTUU /tmp/tmp.X2FNb8sBd7 ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qw7a1F1knN +++ mktemp ++ local LAST_ERR=/tmp/tmp.iYQneDsQRx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qw7a1F1knN ++ cat /tmp/tmp.iYQneDsQRx ++ rm /tmp/tmp.qw7a1F1knN /tmp/tmp.iYQneDsQRx ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.Iojqr3GPLG/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.u4mv5pqwTD +++ mktemp ++ local LAST_ERR=/tmp/tmp.ziwekeLmbp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.u4mv5pqwTD ++ cat /tmp/tmp.ziwekeLmbp ++ rm /tmp/tmp.u4mv5pqwTD /tmp/tmp.ziwekeLmbp ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-cwmpz ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.Iojqr3GPLG/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.blXTKwxKDb +++ mktemp ++ local LAST_ERR=/tmp/tmp.E7AMmJhGV8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.blXTKwxKDb ++ cat /tmp/tmp.E7AMmJhGV8 ++ rm /tmp/tmp.blXTKwxKDb /tmp/tmp.E7AMmJhGV8 ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.Iojqr3GPLG/select-1.sql + is_keyring_plugin_in_use pitr + local cluster=pitr + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + grep -E -o 'early-plugin-load=keyring_\w+.so' + kubectl exec pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + return 1 ++ kubectl_bin get pods --selector=app.kubernetes.io/component=pitr -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HjyBn7G3q0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WFCg3Ay94c ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/component=pitr -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HjyBn7G3q0 ++ cat /tmp/tmp.WFCg3Ay94c ++ rm /tmp/tmp.HjyBn7G3q0 /tmp/tmp.WFCg3Ay94c ++ return 0 + pitr_pod=pitr-pitr-7f6d788949-b6fbt + wait_pod pitr-pitr-7f6d788949-b6fbt + local pod=pitr-pitr-7f6d788949-b6fbt + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pitr-7f6d788949-b6fbt ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pitr-pitr-7f6d788949-b6fbt condition met waiting for pod/pitr-pitr-7f6d788949-b6fbt to become ReadyDefaulted container "pitr" out of: pitr, pitr-init (init) .Ok ++ kubectl_bin logs -l app.kubernetes.io/component=pitr --tail=-1 ++ egrep -c 'already exists|is already installed' + [[ 6 == \0 ]] + run_backup pitr on-pitr-minio + local cluster=pitr + local backup=on-pitr-minio + local name= + '[' -z ']' + name=on-pitr-minio + log 'run backup pxc-backup/on-pitr-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:41:42+0000]' run backup pxc-backup/on-pitr-minio [2026-04-29T16:41:42+0000] run backup pxc-backup/on-pitr-minio + yq eval '.metadata.name="on-pitr-minio"' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/on-pitr-minio.yml ++ mktemp + local LAST_OUT=/tmp/tmp.imyeXxyaR7 ++ mktemp + local LAST_ERR=/tmp/tmp.ocSO0nohVL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.imyeXxyaR7 perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio created + cat /tmp/tmp.ocSO0nohVL + rm /tmp/tmp.imyeXxyaR7 /tmp/tmp.ocSO0nohVL + return 0 + wait_backup on-pitr-minio + local backup=on-pitr-minio + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio to reach Succeeded state.............Succeeded + write_test_data pitr + local cluster=pitr + local config= + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ErKWWj2iEt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yKFSgB2jeK +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.ErKWWj2iEt +++ cat /tmp/tmp.yKFSgB2jeK +++ rm /tmp/tmp.ErKWWj2iEt /tmp/tmp.yKFSgB2jeK +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tB35BLgmLh ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0kI9dAHu0F +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.tB35BLgmLh +++ cat /tmp/tmp.0kI9dAHu0F +++ rm /tmp/tmp.tB35BLgmLh /tmp/tmp.0kI9dAHu0F +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + desc 'write test data' + set +o xtrace ----------------------------------------------------------------------------------- write test data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LDexLRN591 +++ mktemp ++ local LAST_ERR=/tmp/tmp.oGukBYzwzv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LDexLRN591 ++ cat /tmp/tmp.oGukBYzwzv ++ rm /tmp/tmp.LDexLRN591 /tmp/tmp.oGukBYzwzv ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' '-h pitr-proxysql -uroot -proot_password' + local 'command=INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9H3GJYBULC +++ mktemp ++ local LAST_ERR=/tmp/tmp.oFDznWXbja ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9H3GJYBULC ++ cat /tmp/tmp.oFDznWXbja ++ rm /tmp/tmp.9H3GJYBULC /tmp/tmp.oFDznWXbja ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep_with_log 30 + local d=30 + log 'sleeping for 30 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:42:52+0000]' sleeping for 30 seconds [2026-04-29T16:42:52+0000] sleeping for 30 seconds + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k18gNBP21K +++ mktemp ++ local LAST_ERR=/tmp/tmp.g3a49oDnkU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k18gNBP21K ++ cat /tmp/tmp.g3a49oDnkU ++ rm /tmp/tmp.k18gNBP21K /tmp/tmp.g3a49oDnkU ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.Iojqr3GPLG/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Afs4lIRJ9K +++ mktemp ++ local LAST_ERR=/tmp/tmp.G3q3nVQSfV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Afs4lIRJ9K ++ cat /tmp/tmp.G3q3nVQSfV ++ rm /tmp/tmp.Afs4lIRJ9K /tmp/tmp.G3q3nVQSfV ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.Iojqr3GPLG/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OD2ngJ5hrV +++ mktemp ++ local LAST_ERR=/tmp/tmp.srwPqg8ftw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OD2ngJ5hrV ++ cat /tmp/tmp.srwPqg8ftw ++ rm /tmp/tmp.OD2ngJ5hrV /tmp/tmp.srwPqg8ftw ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-cwmpz + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.Iojqr3GPLG/select-2.sql ++ is_keyring_plugin_in_use pitr ++ local cluster=pitr ++ [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] ++ kubectl exec pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ grep -E -o 'early-plugin-load=keyring_\w+.so' ++ return 1 + '[' '' ']' + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass1 + patch_secret my-cluster-secrets xtrabackup cGFzczE= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczE= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.Xdyf7zEJEH ++ mktemp + local LAST_ERR=/tmp/tmp.tzt0e6A3kN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Xdyf7zEJEH secret/my-cluster-secrets patched + cat /tmp/tmp.tzt0e6A3kN + rm /tmp/tmp.Xdyf7zEJEH /tmp/tmp.tzt0e6A3kN + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CuTXjbuoys +++ mktemp ++ local LAST_ERR=/tmp/tmp.PVNd7ryOVi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CuTXjbuoys ++ cat /tmp/tmp.PVNd7ryOVi ++ rm /tmp/tmp.CuTXjbuoys /tmp/tmp.PVNd7ryOVi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cESSjridjU +++ mktemp ++ local LAST_ERR=/tmp/tmp.aZMe4hAPUF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cESSjridjU ++ cat /tmp/tmp.aZMe4hAPUF ++ rm /tmp/tmp.cESSjridjU /tmp/tmp.aZMe4hAPUF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tUVGUGEHjQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.1tIuXJzezQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tUVGUGEHjQ ++ cat /tmp/tmp.1tIuXJzezQ ++ rm /tmp/tmp.tUVGUGEHjQ /tmp/tmp.1tIuXJzezQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3UiiBz3kSW +++ mktemp ++ local LAST_ERR=/tmp/tmp.yXm5W10Zcm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3UiiBz3kSW ++ cat /tmp/tmp.yXm5W10Zcm ++ rm /tmp/tmp.3UiiBz3kSW /tmp/tmp.yXm5W10Zcm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FMIFfJq8Zh +++ mktemp ++ local LAST_ERR=/tmp/tmp.YLNcgvF4jJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FMIFfJq8Zh ++ cat /tmp/tmp.YLNcgvF4jJ ++ rm /tmp/tmp.FMIFfJq8Zh /tmp/tmp.YLNcgvF4jJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3AqSt3c4DZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZzRTTgAzFW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3AqSt3c4DZ ++ cat /tmp/tmp.ZzRTTgAzFW ++ rm /tmp/tmp.3AqSt3c4DZ /tmp/tmp.ZzRTTgAzFW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VqH4Gil6ww +++ mktemp ++ local LAST_ERR=/tmp/tmp.tT4Sbm8itw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VqH4Gil6ww ++ cat /tmp/tmp.tT4Sbm8itw ++ rm /tmp/tmp.VqH4Gil6ww /tmp/tmp.tT4Sbm8itw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hFIiuWzXCv +++ mktemp ++ local LAST_ERR=/tmp/tmp.5gsi6zHFjB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hFIiuWzXCv ++ cat /tmp/tmp.5gsi6zHFjB ++ rm /tmp/tmp.hFIiuWzXCv /tmp/tmp.5gsi6zHFjB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RJRTTc5EZb +++ mktemp ++ local LAST_ERR=/tmp/tmp.3OjAmTdUSB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RJRTTc5EZb ++ cat /tmp/tmp.3OjAmTdUSB ++ rm /tmp/tmp.RJRTTc5EZb /tmp/tmp.3OjAmTdUSB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SXfrZjyQwx +++ mktemp ++ local LAST_ERR=/tmp/tmp.ptUtxzUOKH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SXfrZjyQwx ++ cat /tmp/tmp.ptUtxzUOKH ++ rm /tmp/tmp.SXfrZjyQwx /tmp/tmp.ptUtxzUOKH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DF2Uabbk6Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.dy25v8rqxW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DF2Uabbk6Q ++ cat /tmp/tmp.dy25v8rqxW ++ rm /tmp/tmp.DF2Uabbk6Q /tmp/tmp.dy25v8rqxW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cKPKsKmbR4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FiHxpa4kPJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cKPKsKmbR4 ++ cat /tmp/tmp.FiHxpa4kPJ ++ rm /tmp/tmp.cKPKsKmbR4 /tmp/tmp.FiHxpa4kPJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zksnxFwOP2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SgR1XIICEz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zksnxFwOP2 ++ cat /tmp/tmp.SgR1XIICEz ++ rm /tmp/tmp.zksnxFwOP2 /tmp/tmp.SgR1XIICEz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gks1tGNC5G +++ mktemp ++ local LAST_ERR=/tmp/tmp.M807ckq7MX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gks1tGNC5G ++ cat /tmp/tmp.M807ckq7MX ++ rm /tmp/tmp.gks1tGNC5G /tmp/tmp.M807ckq7MX ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0AsVaLMmmo +++ mktemp ++ local LAST_ERR=/tmp/tmp.YM7z9t6mlq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0AsVaLMmmo ++ cat /tmp/tmp.YM7z9t6mlq ++ rm /tmp/tmp.0AsVaLMmmo /tmp/tmp.YM7z9t6mlq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WGsvrlWYJY +++ mktemp ++ local LAST_ERR=/tmp/tmp.KXo5zmUkLr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WGsvrlWYJY ++ cat /tmp/tmp.KXo5zmUkLr ++ rm /tmp/tmp.WGsvrlWYJY /tmp/tmp.KXo5zmUkLr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1X09VDT2DO +++ mktemp ++ local LAST_ERR=/tmp/tmp.SPPCkoIOoI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1X09VDT2DO ++ cat /tmp/tmp.SPPCkoIOoI ++ rm /tmp/tmp.1X09VDT2DO /tmp/tmp.SPPCkoIOoI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6T6OpJmceQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.FVLSqFQ3dR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6T6OpJmceQ ++ cat /tmp/tmp.FVLSqFQ3dR ++ rm /tmp/tmp.6T6OpJmceQ /tmp/tmp.FVLSqFQ3dR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3hyfm19T1Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wa8o4xw2jk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3hyfm19T1Y ++ cat /tmp/tmp.Wa8o4xw2jk ++ rm /tmp/tmp.3hyfm19T1Y /tmp/tmp.Wa8o4xw2jk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.36g4jqsHbV +++ mktemp ++ local LAST_ERR=/tmp/tmp.rqZnaytbos ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.36g4jqsHbV ++ cat /tmp/tmp.rqZnaytbos ++ rm /tmp/tmp.36g4jqsHbV /tmp/tmp.rqZnaytbos ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Jugd2Bl6w6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.KGiAjcQjRk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Jugd2Bl6w6 ++ cat /tmp/tmp.KGiAjcQjRk ++ rm /tmp/tmp.Jugd2Bl6w6 /tmp/tmp.KGiAjcQjRk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ni3hWT8SVT +++ mktemp ++ local LAST_ERR=/tmp/tmp.geelnjBYsE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ni3hWT8SVT ++ cat /tmp/tmp.geelnjBYsE ++ rm /tmp/tmp.ni3hWT8SVT /tmp/tmp.geelnjBYsE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.En1RyrERth +++ mktemp ++ local LAST_ERR=/tmp/tmp.nkQlpPe4E2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.En1RyrERth ++ cat /tmp/tmp.nkQlpPe4E2 ++ rm /tmp/tmp.En1RyrERth /tmp/tmp.nkQlpPe4E2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.X8gl9zMrXP +++ mktemp ++ local LAST_ERR=/tmp/tmp.JjCvbJ06xC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.X8gl9zMrXP ++ cat /tmp/tmp.JjCvbJ06xC ++ rm /tmp/tmp.X8gl9zMrXP /tmp/tmp.JjCvbJ06xC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PM33Ip9rHG +++ mktemp ++ local LAST_ERR=/tmp/tmp.A1KvDS41ds ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PM33Ip9rHG ++ cat /tmp/tmp.A1KvDS41ds ++ rm /tmp/tmp.PM33Ip9rHG /tmp/tmp.A1KvDS41ds ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oJjSOerrpC +++ mktemp ++ local LAST_ERR=/tmp/tmp.ujyy37hpt2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oJjSOerrpC ++ cat /tmp/tmp.ujyy37hpt2 ++ rm /tmp/tmp.oJjSOerrpC /tmp/tmp.ujyy37hpt2 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.4XvRcoN6J4 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.KxH2CwaHqD +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.4XvRcoN6J4 +++++ cat /tmp/tmp.KxH2CwaHqD +++++ rm /tmp/tmp.4XvRcoN6J4 /tmp/tmp.KxH2CwaHqD +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.FfMt1I4Wrj ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.FTn4c69nCO +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.FfMt1I4Wrj +++++ cat /tmp/tmp.FTn4c69nCO +++++ rm /tmp/tmp.FfMt1I4Wrj /tmp/tmp.FTn4c69nCO +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EJ9Xp9Lh8T +++ mktemp ++ local LAST_ERR=/tmp/tmp.pvx5JhrrOE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EJ9Xp9Lh8T ++ cat /tmp/tmp.pvx5JhrrOE ++ rm /tmp/tmp.EJ9Xp9Lh8T /tmp/tmp.pvx5JhrrOE ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass2 + patch_secret my-cluster-secrets xtrabackup cGFzczI= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.CluVQec7QI ++ mktemp + local LAST_ERR=/tmp/tmp.LOSu4JqZEC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CluVQec7QI secret/my-cluster-secrets patched + cat /tmp/tmp.LOSu4JqZEC + rm /tmp/tmp.CluVQec7QI /tmp/tmp.LOSu4JqZEC + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fVpo8InZR9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SDcgnhWUCK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fVpo8InZR9 ++ cat /tmp/tmp.SDcgnhWUCK ++ rm /tmp/tmp.fVpo8InZR9 /tmp/tmp.SDcgnhWUCK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Mozet1LhDh +++ mktemp ++ local LAST_ERR=/tmp/tmp.BvpZNneHyA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Mozet1LhDh ++ cat /tmp/tmp.BvpZNneHyA ++ rm /tmp/tmp.Mozet1LhDh /tmp/tmp.BvpZNneHyA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h3cwgOg7gr +++ mktemp ++ local LAST_ERR=/tmp/tmp.yk8fLPNyBo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h3cwgOg7gr ++ cat /tmp/tmp.yk8fLPNyBo ++ rm /tmp/tmp.h3cwgOg7gr /tmp/tmp.yk8fLPNyBo ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cBjkxbxoMD +++ mktemp ++ local LAST_ERR=/tmp/tmp.8JBqzj6uFe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cBjkxbxoMD ++ cat /tmp/tmp.8JBqzj6uFe ++ rm /tmp/tmp.cBjkxbxoMD /tmp/tmp.8JBqzj6uFe ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lC1isAUnIc +++ mktemp ++ local LAST_ERR=/tmp/tmp.XStE47ARW8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lC1isAUnIc ++ cat /tmp/tmp.XStE47ARW8 ++ rm /tmp/tmp.lC1isAUnIc /tmp/tmp.XStE47ARW8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x0W4ImKjFB +++ mktemp ++ local LAST_ERR=/tmp/tmp.wGCCA9rI1X ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.x0W4ImKjFB ++ cat /tmp/tmp.wGCCA9rI1X ++ rm /tmp/tmp.x0W4ImKjFB /tmp/tmp.wGCCA9rI1X ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XCP8r2VBIV +++ mktemp ++ local LAST_ERR=/tmp/tmp.dTUbXnMCqV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XCP8r2VBIV ++ cat /tmp/tmp.dTUbXnMCqV ++ rm /tmp/tmp.XCP8r2VBIV /tmp/tmp.dTUbXnMCqV ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xIAa4F2xEu +++ mktemp ++ local LAST_ERR=/tmp/tmp.HhJPQ1O374 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xIAa4F2xEu ++ cat /tmp/tmp.HhJPQ1O374 ++ rm /tmp/tmp.xIAa4F2xEu /tmp/tmp.HhJPQ1O374 ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BvVTd3V9Cs +++ mktemp ++ local LAST_ERR=/tmp/tmp.79Y9LOFLAP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BvVTd3V9Cs ++ cat /tmp/tmp.79Y9LOFLAP ++ rm /tmp/tmp.BvVTd3V9Cs /tmp/tmp.79Y9LOFLAP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vpfp35VYt0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.lUbTiqr99Q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vpfp35VYt0 ++ cat /tmp/tmp.lUbTiqr99Q ++ rm /tmp/tmp.vpfp35VYt0 /tmp/tmp.lUbTiqr99Q ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xx1WdSNxk0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LkBwqPtVIl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xx1WdSNxk0 ++ cat /tmp/tmp.LkBwqPtVIl ++ rm /tmp/tmp.Xx1WdSNxk0 /tmp/tmp.LkBwqPtVIl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VGCpDygvQO +++ mktemp ++ local LAST_ERR=/tmp/tmp.hrcwFGtnvg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VGCpDygvQO ++ cat /tmp/tmp.hrcwFGtnvg ++ rm /tmp/tmp.VGCpDygvQO /tmp/tmp.hrcwFGtnvg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oyhxZD1EIP +++ mktemp ++ local LAST_ERR=/tmp/tmp.yj6n0i2Xqk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oyhxZD1EIP ++ cat /tmp/tmp.yj6n0i2Xqk ++ rm /tmp/tmp.oyhxZD1EIP /tmp/tmp.yj6n0i2Xqk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MwqOhNNPsV +++ mktemp ++ local LAST_ERR=/tmp/tmp.6pQetXBd68 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MwqOhNNPsV ++ cat /tmp/tmp.6pQetXBd68 ++ rm /tmp/tmp.MwqOhNNPsV /tmp/tmp.6pQetXBd68 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I8GHQRWZRp +++ mktemp ++ local LAST_ERR=/tmp/tmp.S32Rq1s5RU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.I8GHQRWZRp ++ cat /tmp/tmp.S32Rq1s5RU ++ rm /tmp/tmp.I8GHQRWZRp /tmp/tmp.S32Rq1s5RU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tEQ4o8ybWe +++ mktemp ++ local LAST_ERR=/tmp/tmp.psLNvFaHvT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tEQ4o8ybWe ++ cat /tmp/tmp.psLNvFaHvT ++ rm /tmp/tmp.tEQ4o8ybWe /tmp/tmp.psLNvFaHvT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xJDYIDh8TU +++ mktemp ++ local LAST_ERR=/tmp/tmp.U1NYARYTAJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xJDYIDh8TU ++ cat /tmp/tmp.U1NYARYTAJ ++ rm /tmp/tmp.xJDYIDh8TU /tmp/tmp.U1NYARYTAJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q60VFGQVJZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.prr2vHUxMH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.q60VFGQVJZ ++ cat /tmp/tmp.prr2vHUxMH ++ rm /tmp/tmp.q60VFGQVJZ /tmp/tmp.prr2vHUxMH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S49aj8cfCD +++ mktemp ++ local LAST_ERR=/tmp/tmp.5MKGSuKL7u ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S49aj8cfCD ++ cat /tmp/tmp.5MKGSuKL7u ++ rm /tmp/tmp.S49aj8cfCD /tmp/tmp.5MKGSuKL7u ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y8DDhmHLaU +++ mktemp ++ local LAST_ERR=/tmp/tmp.AVp10qJqRP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y8DDhmHLaU ++ cat /tmp/tmp.AVp10qJqRP ++ rm /tmp/tmp.Y8DDhmHLaU /tmp/tmp.AVp10qJqRP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xk7n3dyQAY +++ mktemp ++ local LAST_ERR=/tmp/tmp.GiMgs5lPrV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xk7n3dyQAY ++ cat /tmp/tmp.GiMgs5lPrV ++ rm /tmp/tmp.Xk7n3dyQAY /tmp/tmp.GiMgs5lPrV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2PFETjledi +++ mktemp ++ local LAST_ERR=/tmp/tmp.qiJPH7IXm8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2PFETjledi ++ cat /tmp/tmp.qiJPH7IXm8 ++ rm /tmp/tmp.2PFETjledi /tmp/tmp.qiJPH7IXm8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MvLVfEGyTl +++ mktemp ++ local LAST_ERR=/tmp/tmp.rAJHUZRWY0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MvLVfEGyTl ++ cat /tmp/tmp.rAJHUZRWY0 ++ rm /tmp/tmp.MvLVfEGyTl /tmp/tmp.rAJHUZRWY0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.O8hPtFScSa +++ mktemp ++ local LAST_ERR=/tmp/tmp.LxevnzYstt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.O8hPtFScSa ++ cat /tmp/tmp.LxevnzYstt ++ rm /tmp/tmp.O8hPtFScSa /tmp/tmp.LxevnzYstt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iR96N32Fc9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.U7whxAAWJs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iR96N32Fc9 ++ cat /tmp/tmp.U7whxAAWJs ++ rm /tmp/tmp.iR96N32Fc9 /tmp/tmp.U7whxAAWJs ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ksCVFKkokB +++ mktemp ++ local LAST_ERR=/tmp/tmp.NzPpSWrZK2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ksCVFKkokB ++ cat /tmp/tmp.NzPpSWrZK2 ++ rm /tmp/tmp.ksCVFKkokB /tmp/tmp.NzPpSWrZK2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2oz4VISgmD +++ mktemp ++ local LAST_ERR=/tmp/tmp.YjvVCpfIGZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2oz4VISgmD ++ cat /tmp/tmp.YjvVCpfIGZ ++ rm /tmp/tmp.2oz4VISgmD /tmp/tmp.YjvVCpfIGZ ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GGwBJHrVzy +++ mktemp ++ local LAST_ERR=/tmp/tmp.HD74KLjSgo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GGwBJHrVzy ++ cat /tmp/tmp.HD74KLjSgo ++ rm /tmp/tmp.GGwBJHrVzy /tmp/tmp.HD74KLjSgo ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.5XhN4qskKA ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.vI3lweij8N +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.5XhN4qskKA +++++ cat /tmp/tmp.vI3lweij8N +++++ rm /tmp/tmp.5XhN4qskKA /tmp/tmp.vI3lweij8N +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.bC2jRs40i8 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.XiVgvrNp6C +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.bC2jRs40i8 +++++ cat /tmp/tmp.XiVgvrNp6C +++++ rm /tmp/tmp.bC2jRs40i8 /tmp/tmp.XiVgvrNp6C +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rwagAd1Jms +++ mktemp ++ local LAST_ERR=/tmp/tmp.RjIvTAr78R ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rwagAd1Jms ++ cat /tmp/tmp.RjIvTAr78R ++ rm /tmp/tmp.rwagAd1Jms /tmp/tmp.RjIvTAr78R ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ echo -n pass3 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczM= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczM= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.uiTbsYjrZG ++ mktemp + local LAST_ERR=/tmp/tmp.qzqvywpgco + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uiTbsYjrZG secret/my-cluster-secrets patched + cat /tmp/tmp.qzqvywpgco + rm /tmp/tmp.uiTbsYjrZG /tmp/tmp.qzqvywpgco + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mqHzH3hGIc +++ mktemp ++ local LAST_ERR=/tmp/tmp.s8Jruze8vV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mqHzH3hGIc ++ cat /tmp/tmp.s8Jruze8vV ++ rm /tmp/tmp.mqHzH3hGIc /tmp/tmp.s8Jruze8vV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kKi0KpvzIu +++ mktemp ++ local LAST_ERR=/tmp/tmp.eDMlhetrUG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kKi0KpvzIu ++ cat /tmp/tmp.eDMlhetrUG ++ rm /tmp/tmp.kKi0KpvzIu /tmp/tmp.eDMlhetrUG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xniSWNc69l +++ mktemp ++ local LAST_ERR=/tmp/tmp.g0V7YNUeQx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xniSWNc69l ++ cat /tmp/tmp.g0V7YNUeQx ++ rm /tmp/tmp.xniSWNc69l /tmp/tmp.g0V7YNUeQx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tosLbyKgus +++ mktemp ++ local LAST_ERR=/tmp/tmp.7YPWVKOOgA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tosLbyKgus ++ cat /tmp/tmp.7YPWVKOOgA ++ rm /tmp/tmp.tosLbyKgus /tmp/tmp.7YPWVKOOgA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WwMtbzn0po +++ mktemp ++ local LAST_ERR=/tmp/tmp.nmzNj0WPPu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WwMtbzn0po ++ cat /tmp/tmp.nmzNj0WPPu ++ rm /tmp/tmp.WwMtbzn0po /tmp/tmp.nmzNj0WPPu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5eRPMgmXt4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.RHTejMi0XD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5eRPMgmXt4 ++ cat /tmp/tmp.RHTejMi0XD ++ rm /tmp/tmp.5eRPMgmXt4 /tmp/tmp.RHTejMi0XD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AGqAZjPrxU +++ mktemp ++ local LAST_ERR=/tmp/tmp.TLcY21TBQw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AGqAZjPrxU ++ cat /tmp/tmp.TLcY21TBQw ++ rm /tmp/tmp.AGqAZjPrxU /tmp/tmp.TLcY21TBQw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0sfs2zI1wL +++ mktemp ++ local LAST_ERR=/tmp/tmp.PtgZ5Dww2b ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0sfs2zI1wL ++ cat /tmp/tmp.PtgZ5Dww2b ++ rm /tmp/tmp.0sfs2zI1wL /tmp/tmp.PtgZ5Dww2b ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y3iGT8e6PF +++ mktemp ++ local LAST_ERR=/tmp/tmp.BvnHFAwnNo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y3iGT8e6PF ++ cat /tmp/tmp.BvnHFAwnNo ++ rm /tmp/tmp.Y3iGT8e6PF /tmp/tmp.BvnHFAwnNo ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uVK5baPSBX +++ mktemp ++ local LAST_ERR=/tmp/tmp.6Af7gGnHAG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uVK5baPSBX ++ cat /tmp/tmp.6Af7gGnHAG ++ rm /tmp/tmp.uVK5baPSBX /tmp/tmp.6Af7gGnHAG ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aqh0eUJwIj +++ mktemp ++ local LAST_ERR=/tmp/tmp.qYvA0Jp8iH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aqh0eUJwIj ++ cat /tmp/tmp.qYvA0Jp8iH ++ rm /tmp/tmp.aqh0eUJwIj /tmp/tmp.qYvA0Jp8iH ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NC8JUq1IMr +++ mktemp ++ local LAST_ERR=/tmp/tmp.wy8lq8wTM2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NC8JUq1IMr ++ cat /tmp/tmp.wy8lq8wTM2 ++ rm /tmp/tmp.NC8JUq1IMr /tmp/tmp.wy8lq8wTM2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SKgCCNn5oJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.rk0lTROz3n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SKgCCNn5oJ ++ cat /tmp/tmp.rk0lTROz3n ++ rm /tmp/tmp.SKgCCNn5oJ /tmp/tmp.rk0lTROz3n ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fFyfpBngNs +++ mktemp ++ local LAST_ERR=/tmp/tmp.yuxrwtd3Hy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fFyfpBngNs ++ cat /tmp/tmp.yuxrwtd3Hy ++ rm /tmp/tmp.fFyfpBngNs /tmp/tmp.yuxrwtd3Hy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v9H4aaK55E +++ mktemp ++ local LAST_ERR=/tmp/tmp.aZEMob6wBF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v9H4aaK55E ++ cat /tmp/tmp.aZEMob6wBF ++ rm /tmp/tmp.v9H4aaK55E /tmp/tmp.aZEMob6wBF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.321dsCBeRN +++ mktemp ++ local LAST_ERR=/tmp/tmp.PEFxRldouW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.321dsCBeRN ++ cat /tmp/tmp.PEFxRldouW ++ rm /tmp/tmp.321dsCBeRN /tmp/tmp.PEFxRldouW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lLQnZux1p5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.f5fHfhP4GH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lLQnZux1p5 ++ cat /tmp/tmp.f5fHfhP4GH ++ rm /tmp/tmp.lLQnZux1p5 /tmp/tmp.f5fHfhP4GH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FoUOZtUKdz +++ mktemp ++ local LAST_ERR=/tmp/tmp.CvHea8QgnY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FoUOZtUKdz ++ cat /tmp/tmp.CvHea8QgnY ++ rm /tmp/tmp.FoUOZtUKdz /tmp/tmp.CvHea8QgnY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cf7NxJD0gY +++ mktemp ++ local LAST_ERR=/tmp/tmp.oTRuvxpb9d ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cf7NxJD0gY ++ cat /tmp/tmp.oTRuvxpb9d ++ rm /tmp/tmp.cf7NxJD0gY /tmp/tmp.oTRuvxpb9d ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.javiu4gqmw +++ mktemp ++ local LAST_ERR=/tmp/tmp.3iab7EiHz5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.javiu4gqmw ++ cat /tmp/tmp.3iab7EiHz5 ++ rm /tmp/tmp.javiu4gqmw /tmp/tmp.3iab7EiHz5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9MW4Vz139F +++ mktemp ++ local LAST_ERR=/tmp/tmp.3m8OQhc37x ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9MW4Vz139F ++ cat /tmp/tmp.3m8OQhc37x ++ rm /tmp/tmp.9MW4Vz139F /tmp/tmp.3m8OQhc37x ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R96mRP8Wg9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.6L7un9773e ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.R96mRP8Wg9 ++ cat /tmp/tmp.6L7un9773e ++ rm /tmp/tmp.R96mRP8Wg9 /tmp/tmp.6L7un9773e ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.O3njcRfz7y +++ mktemp ++ local LAST_ERR=/tmp/tmp.05gt65rblt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.O3njcRfz7y ++ cat /tmp/tmp.05gt65rblt ++ rm /tmp/tmp.O3njcRfz7y /tmp/tmp.05gt65rblt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U60lZFwZor +++ mktemp ++ local LAST_ERR=/tmp/tmp.mikHKFYdKx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.U60lZFwZor ++ cat /tmp/tmp.mikHKFYdKx ++ rm /tmp/tmp.U60lZFwZor /tmp/tmp.mikHKFYdKx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IAcy910WAH +++ mktemp ++ local LAST_ERR=/tmp/tmp.W0gl2qoP36 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IAcy910WAH ++ cat /tmp/tmp.W0gl2qoP36 ++ rm /tmp/tmp.IAcy910WAH /tmp/tmp.W0gl2qoP36 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ks5zt22Ua5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ir8JFaVtN6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ks5zt22Ua5 ++ cat /tmp/tmp.ir8JFaVtN6 ++ rm /tmp/tmp.ks5zt22Ua5 /tmp/tmp.ir8JFaVtN6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.54fAOn0MBO +++ mktemp ++ local LAST_ERR=/tmp/tmp.47vBut1Mc4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.54fAOn0MBO ++ cat /tmp/tmp.47vBut1Mc4 ++ rm /tmp/tmp.54fAOn0MBO /tmp/tmp.47vBut1Mc4 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zAoxBCE0iQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.o3bC1eoRk6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zAoxBCE0iQ ++ cat /tmp/tmp.o3bC1eoRk6 ++ rm /tmp/tmp.zAoxBCE0iQ /tmp/tmp.o3bC1eoRk6 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.KF6iOJMBX9 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.jlR5P73DI1 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.KF6iOJMBX9 +++++ cat /tmp/tmp.jlR5P73DI1 +++++ rm /tmp/tmp.KF6iOJMBX9 /tmp/tmp.jlR5P73DI1 +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.gh2IjVumQs ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.VO0v1mKK1e +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.gh2IjVumQs +++++ cat /tmp/tmp.VO0v1mKK1e +++++ rm /tmp/tmp.gh2IjVumQs /tmp/tmp.VO0v1mKK1e +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Sz6qJxdvN4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.YWDeP7a333 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Sz6qJxdvN4 ++ cat /tmp/tmp.YWDeP7a333 ++ rm /tmp/tmp.Sz6qJxdvN4 /tmp/tmp.YWDeP7a333 ++ return 0 + [[ 2 == \2 ]] + echo + compare_kubectl secret/pitr-mysql-init + local resource=secret/pitr-mysql-init + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml + local new_result=/tmp/tmp.Iojqr3GPLG/secret_pitr-mysql-init.yml + desc 'compare secret/pitr-mysql-init-' + set +o xtrace ----------------------------------------------------------------------------------- compare secret/pitr-mysql-init- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129.yml ']' + version_gt 1.27 ++ echo '1.32 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k127.yml ']' + version_gt 1.24 ++ echo '1.32 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k124.yml ']' + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k122.yml ']' + version_gt 1.21 ++ echo '1.32 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init-aks.yml ']' + kubectl_bin get -o yaml secret/pitr-mysql-init ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("pitr-18261", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.M3bAl5eWRD ++ mktemp + local LAST_ERR=/tmp/tmp.00a1U8pDIN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml secret/pitr-mysql-init + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.M3bAl5eWRD + cat /tmp/tmp.00a1U8pDIN + rm /tmp/tmp.M3bAl5eWRD /tmp/tmp.00a1U8pDIN + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml /tmp/tmp.Iojqr3GPLG/secret_pitr-mysql-init.yml + log 'compare_kubectl: secret/pitr-mysql-init OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:53:37+0000]' compare_kubectl: secret/pitr-mysql-init OK [2026-04-29T16:53:37+0000] compare_kubectl: secret/pitr-mysql-init OK + mysql_host=pitr-pxc-0.pitr-pxc.pitr-18261 ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-18261 ++ local host=pitr-pxc-0.pitr-pxc.pitr-18261 +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.nk8jj1Ia7y +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.DwldJzpD3J ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.nk8jj1Ia7y ++++ cat /tmp/tmp.DwldJzpD3J ++++ rm /tmp/tmp.nk8jj1Ia7y /tmp/tmp.DwldJzpD3J ++++ return 0 +++ client_pod=pxc-client-67fc4995bb-cwmpz +++ wait_pod pxc-client-67fc4995bb-cwmpz +++ local pod=pxc-client-67fc4995bb-cwmpz +++ local max_retry=480 +++ local ns= ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ grep -E '^(pxc|proxysql)$' ++++ echo pxc-client-67fc4995bb-cwmpz +++ local container= +++ set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local gtid=728a337c-43e9-11f1-ae89-223d0a56030a:29 ++ [[ ! 728a337c-43e9-11f1-ae89-223d0a56030a:29 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo 728a337c-43e9-11f1-ae89-223d0a56030a:29 + gtid=728a337c-43e9-11f1-ae89-223d0a56030a:29 + log '[CASE 1]: gtid_executed: 728a337c-43e9-11f1-ae89-223d0a56030a:29' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:53:47+0000]' '[CASE' '1]:' gtid_executed: 728a337c-43e9-11f1-ae89-223d0a56030a:29 [2026-04-29T16:53:47+0000] [CASE 1]: gtid_executed: 728a337c-43e9-11f1-ae89-223d0a56030a:29 + write_data_for_pitr pitr 100503 3 + local cluster=pitr + local start=100503 + local rows=3 ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.azpWfFUftS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.U2B2lijQpy +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.azpWfFUftS +++ cat /tmp/tmp.U2B2lijQpy +++ rm /tmp/tmp.azpWfFUftS /tmp/tmp.U2B2lijQpy +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OeTD0uNxrD ++++ mktemp +++ local LAST_ERR=/tmp/tmp.yXEOPkeXF3 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.OeTD0uNxrD +++ cat /tmp/tmp.yXEOPkeXF3 +++ rm /tmp/tmp.OeTD0uNxrD /tmp/tmp.yXEOPkeXF3 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + local sql= + (( i=0 )) + (( i//g' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-18261~ + kubectl_bin apply -f - + /usr/bin/sed -e 's//728a337c-43e9-11f1-ae89-223d0a56030a:29/g' + /usr/bin/sed -e 's///g' ++ mktemp + local LAST_OUT=/tmp/tmp.EPb2SuxI1m ++ mktemp + local LAST_ERR=/tmp/tmp.6m8zqCMYyM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EPb2SuxI1m perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-gtid created + cat /tmp/tmp.6m8zqCMYyM + rm /tmp/tmp.EPb2SuxI1m /tmp/tmp.6m8zqCMYyM + return 0 + wait_backup_restore restore-on-pitr-minio-gtid 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Stopping Cluster state 2026-04-29T16:55:13 pxc-restore/restore-on-pitr-minio-gtid state: Stopping Cluster + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-1" not found + wait_for_delete pod/pitr-pxc-0 + local res=pod/pitr-pxc-0 + echo -n 'waiting for pod/pitr-pxc-0 to be deleted' waiting for pod/pitr-pxc-0 to be deleted+ set +o xtrace ......Error from server (NotFound): pods "pitr-pxc-0" not found + wait_backup_restore restore-on-pitr-minio-gtid 'Point-in-time recovering' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Point-in-time recovering' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Point-in-time recovering state 2026-04-29T16:55:52 pxc-restore/restore-on-pitr-minio-gtid state: Stopping Cluster 2026-04-29T16:55:54 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:55:56 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:55:59 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:01 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:03 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:05 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:08 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:10 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:12 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:14 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:17 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:19 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:22 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:24 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:27 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:29 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:31 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:33 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:36 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:38 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:41 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:44 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:46 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:49 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:51 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:53 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:56 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:56:59 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:01 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:03 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:06 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:08 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:10 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:13 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:15 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:17 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T16:57:20 pxc-restore/restore-on-pitr-minio-gtid state: Point-in-time recovering + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-1" not found + wait_pod pitr-pxc-0 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-pxc-0 + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + wait_backup_restore restore-on-pitr-minio-gtid 'Starting Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Starting Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Starting Cluster state 2026-04-29T16:57:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.txZSGkiicu +++ mktemp ++ local LAST_ERR=/tmp/tmp.MN9F5se2Yd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.txZSGkiicu ++ cat /tmp/tmp.MN9F5se2Yd ++ rm /tmp/tmp.txZSGkiicu /tmp/tmp.MN9F5se2Yd ++ return 0 + local minio_pod=minio-service-757b74c96c-2vzxj + kubectl_bin exec minio-service-757b74c96c-2vzxj -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json ++ mktemp + local LAST_OUT=/tmp/tmp.rQvCdfJliK ++ mktemp + local LAST_ERR=/tmp/tmp.OQeT9VF5JA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec minio-service-757b74c96c-2vzxj -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl exec minio-service-757b74c96c-2vzxj -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl exec minio-service-757b74c96c-2vzxj -- stat /exports/operator-testing/binlogs/gtid-binlog-cache.json + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.rQvCdfJliK + cat /tmp/tmp.OQeT9VF5JA stat: cannot statx '/exports/operator-testing/binlogs/gtid-binlog-cache.json': No such file or directory command terminated with exit code 1 + rm /tmp/tmp.rQvCdfJliK /tmp/tmp.OQeT9VF5JA + return 1 + wait_backup_restore restore-on-pitr-minio-gtid + local backup_name=restore-on-pitr-minio-gtid + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Succeeded state 2026-04-29T16:57:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:57:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:57:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:57:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:57:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:57:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:57:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:57:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:58:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:14 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T16:59:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:00:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:00:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:00:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting Cluster 2026-04-29T17:00:10 pxc-restore/restore-on-pitr-minio-gtid state: Succeeded + kubectl_bin logs job/restore-job-restore-on-pitr-minio-gtid-pitr ++ mktemp + local LAST_OUT=/tmp/tmp.89koEXAKbz ++ mktemp + local LAST_ERR=/tmp/tmp.wbqMix4LpP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs job/restore-job-restore-on-pitr-minio-gtid-pitr + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.89koEXAKbz + LIB_PATH=/opt/percona/backup/lib/pxc + . /opt/percona/backup/lib/pxc/check-version.sh + . /opt/percona/backup/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + . /opt/percona/backup/lib/pxc/aws.sh ++ set -o errexit ++ export AWS_SHARED_CREDENTIALS_FILE=/tmp/aws-credfile ++ AWS_SHARED_CREDENTIALS_FILE=/tmp/aws-credfile ++ export AWS_REGION=us-east-1 ++ AWS_REGION=us-east-1 ++ export AWS_ENDPOINT_URL=https://minio-service.pitr-18261:9000/ ++ AWS_ENDPOINT_URL=https://minio-service.pitr-18261:9000/ ++ '[' -n false ']' ++ [[ false == \f\a\l\s\e ]] ++ AWS_S3_NO_VERIFY_SSL=--no-verify-ssl ++ caBundleDir=/etc/s3/certs ++ caBundleFile=/etc/s3/certs/ca.crt ++ '[' -f /etc/s3/certs/ca.crt ']' + XBCLOUD_ARGS='--curl-retriable-errors=7 ' + '[' -n false ']' + [[ false == \f\a\l\s\e ]] + XBCLOUD_ARGS='--insecure --curl-retriable-errors=7 ' + '[' -n operator-testing/pitr-2026-04-29-16:41:47-full ']' + set -x + aws --no-verify-ssl s3 ls operator-testing/pitr-2026-04-29-16:41:47-full urllib3/connectionpool.py:1097: InsecureRequestWarning: Unverified HTTPS request is being made to host 'minio-service.pitr-18261'. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/latest/advanced-usage.html#tls-warnings PRE pitr-2026-04-29-16:41:47-full.sst_info/ PRE pitr-2026-04-29-16:41:47-full/ 2026-04-29 16:42:14 26760 pitr-2026-04-29-16:41:47-full.md5 2026-04-29 16:42:01 128 pitr-2026-04-29-16:41:47-full.sst_info.md5 + '[' -n '' ']' + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#ib_16384_10.dblwr /datadir/#ib_16384_11.dblwr /datadir/#ib_16384_12.dblwr /datadir/#ib_16384_13.dblwr /datadir/#ib_16384_14.dblwr /datadir/#ib_16384_15.dblwr /datadir/#ib_16384_2.dblwr /datadir/#ib_16384_3.dblwr /datadir/#ib_16384_4.dblwr /datadir/#ib_16384_5.dblwr /datadir/#ib_16384_6.dblwr /datadir/#ib_16384_7.dblwr /datadir/#ib_16384_8.dblwr /datadir/#ib_16384_9.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000001 /datadir/binlog.000002 /datadir/binlog.000003 /datadir/binlog.000004 /datadir/binlog.000005 /datadir/binlog.000006 /datadir/binlog.000007 /datadir/binlog.000008 /datadir/binlog.000009 /datadir/binlog.000010 /datadir/binlog.000011 /datadir/binlog.000012 /datadir/binlog.000013 /datadir/binlog.000014 /datadir/binlog.000015 /datadir/binlog.000016 /datadir/binlog.000017 /datadir/binlog.000018 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql-state-monitor /datadir/mysql-state-monitor.log /datadir/mysql.ibd /datadir/mysql.state /datadir/notify.sock /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/prepare_restored_cluster.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sys /datadir/test /datadir/undo_001 /datadir/undo_002 /datadir/version_info /datadir/wsrep_cmd_notify_handler.sh /datadir/xtrabackup-server-sidecar ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_wgsL ++ grep -c processor /proc/cpuinfo ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_wgsL --parallel=4 ++ destination ++ '[' -n operator-testing/pitr-2026-04-29-16:41:47-full ']' ++ echo -n s3://operator-testing/pitr-2026-04-29-16:41:47-full + xbcloud get --parallel=4 --insecure --curl-retriable-errors=7 s3://operator-testing/pitr-2026-04-29-16:41:47-full.sst_info error: http request failed: Couldn't resolve host name error: http request failed: Couldn't resolve host name 260429 16:56:02 xbcloud: Successfully connected. 260429 16:56:02 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full.sst_info/sst_info.00000000000000000000. 260429 16:56:02 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full.sst_info/sst_info.00000000000000000000, size 145 260429 16:56:02 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full.sst_info/sst_info.00000000000000000001. 260429 16:56:02 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full.sst_info/sst_info.00000000000000000001, size 22 260429 16:56:02 xbcloud: Download completed. ++ get_xtrabackup_version +++ xtrabackup --version +++ awk '{print $3}' +++ sed 's/-.*//' +++ grep '^xtrabackup version' ++ xtrabackup_ver=8.0.35 ++ echo 8.0.35 + XTRABACKUP_VERSION=8.0.35 + check_for_version 8.0.35 8.0.0 + '[' -z 8.0.35 ']' + '[' -z 8.0.0 ']' + local local_version_str + local required_version_str ++ normalize_version 8.0.35 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.35 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=35 ++ printf %02d%02d%02d 8 0 35 + local_version_str=080035 ++ normalize_version 8.0.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=0 ++ printf %02d%02d%02d 8 0 0 + required_version_str=080000 + [[ 080035 < 080000 ]] + return 0 + XBSTREAM_EXTRA_ARGS=' --decompress' ++ grep -c processor /proc/cpuinfo ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_wgsL --parallel=4 --decompress ++ destination ++ '[' -n operator-testing/pitr-2026-04-29-16:41:47-full ']' ++ echo -n s3://operator-testing/pitr-2026-04-29-16:41:47-full + xbcloud get --parallel=4 --insecure --curl-retriable-errors=7 s3://operator-testing/pitr-2026-04-29-16:41:47-full error: http request failed: Couldn't resolve host name error: http request failed: Couldn't resolve host name 260429 16:56:02 xbcloud: Successfully connected. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_tablespaces.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_logfile.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_info.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_binlog_info.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_binlog_info.lz4.00000000000000000000, size 148 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_logfile.lz4.00000000000000000000, size 498 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_tablespaces.lz4.00000000000000000000, size 128 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_info.lz4.00000000000000000000, size 752 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_binlog_info.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_info.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_logfile.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_tablespaces.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_binlog_info.lz4.00000000000000000001, size 40 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_tablespaces.lz4.00000000000000000001, size 40 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_logfile.lz4.00000000000000000001, size 36 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_info.lz4.00000000000000000001, size 33 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/undo_002.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/undo_001.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/sys/sys_config.ibd.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/variables_info_195.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/undo_002.lz4.00000000000000000000, size 149736 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/sys/sys_config.ibd.lz4.00000000000000000000, size 2714 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/variables_info_195.sdi.lz4.00000000000000000000, size 1875 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/undo_001.lz4.00000000000000000000, size 1947875 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/variables_info_195.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/sys/sys_config.ibd.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/undo_001.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/undo_002.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/variables_info_195.sdi.lz4.00000000000000000001, size 59 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/undo_002.lz4.00000000000000000001, size 26 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/sys/sys_config.ibd.lz4.00000000000000000001, size 36 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/undo_001.lz4.00000000000000000001, size 26 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/variables_by_thr_192.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000000, size 1831 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000000, size 1521 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000000, size 1714 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/variables_by_thr_192.sdi.lz4.00000000000000000000, size 1564 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/variables_by_thr_192.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_143.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_144.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_180.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/variables_by_thr_192.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_101.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000000, size 2528 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000000, size 2701 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000000, size 1669 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_101.sdi.lz4.00000000000000000000, size 1519 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_101.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_147.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_101.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_141.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_142.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/backup-my.cnf.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_148.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_181.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/backup-my.cnf.lz4.00000000000000000000, size 430 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000000, size 1996 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_181.sdi.lz4.00000000000000000000, size 1993 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_148.sdi.lz4.00000000000000000000, size 1693 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_181.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_148.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/backup-my.cnf.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_181.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_148.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_135.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/backup-my.cnf.lz4.00000000000000000001, size 31 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_host_187.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000000, size 1689 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000000, size 3065 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_host_187.sdi.lz4.00000000000000000000, size 1466 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000000, size 1767 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_host_187.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_host_187.sdi.lz4.00000000000000000001, size 59 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_129.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_145.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_139.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_128.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/keyring_componen_202.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/persisted_variab_196.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_128.sdi.lz4.00000000000000000000, size 3069 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/keyring_componen_202.sdi.lz4.00000000000000000000, size 1267 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000000, size 1714 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/persisted_variab_196.sdi.lz4.00000000000000000000, size 1446 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/persisted_variab_196.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/keyring_componen_202.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_128.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_asyn_182.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/persisted_variab_196.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/keyring_componen_202.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_128.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000000, size 2269 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000000, size 1610 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000000, size 2275 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000000, size 1670 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_136.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_94.sdi.lz4.00000000000000000001, size 60 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_125.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_124.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/binary_log_trans_198.sdi.lz4.00000000000000000000. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_checkpoints.00000000000000000000. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000000, size 1765 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000000, size 2099 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/binary_log_trans_198.sdi.lz4.00000000000000000000, size 2078 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_checkpoints.00000000000000000000, size 190 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/xtrabackup_checkpoints.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/binary_log_trans_198.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000001. 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/xtrabackup_checkpoints.00000000000000000001, size 36 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_122.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/binary_log_trans_198.sdi.lz4.00000000000000000001, size 61 260429 16:56:03 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_134.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSM.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/binlog.000009.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSM.lz4.00000000000000000000, size 96 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000000, size 1647 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000000, size 1765 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/binlog.000009.lz4.00000000000000000000, size 226 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/binlog.000009.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSM.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/binlog.000009.lz4.00000000000000000001, size 31 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_152.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_150.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSM.lz4.00000000000000000001, size 36 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/file_instances_102.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/objects_summary__107.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_171.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/file_instances_102.sdi.lz4.00000000000000000000, size 1492 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/objects_summary__107.sdi.lz4.00000000000000000000, size 1541 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_171.sdi.lz4.00000000000000000000, size 3075 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000000, size 1605 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_171.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/objects_summary__107.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/file_instances_102.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_171.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/objects_summary__107.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/file_instances_102.sdi.lz4.00000000000000000001, size 59 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_126.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_consumers_112.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/users_153.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000000, size 1927 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_consumers_112.sdi.lz4.00000000000000000000, size 1438 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/users_153.sdi.lz4.00000000000000000000, size 1495 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000000, size 2170 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/users_153.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_consumers_112.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/users_153.sdi.lz4.00000000000000000001, size 50 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_132.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_117.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_consumers_112.sdi.lz4.00000000000000000001, size 60 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/general_log_225.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_121.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__104.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/general_log_225.sdi.lz4.00000000000000000000, size 1489 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__104.sdi.lz4.00000000000000000000, size 1961 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000000, size 1738 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_121.sdi.lz4.00000000000000000000, size 1907 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__104.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_121.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/general_log_225.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__104.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/general_log_225.sdi.lz4.00000000000000000001, size 43 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_hi_121.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_151.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/slow_log_226.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/global_variables_193.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/ibdata1.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/slow_log_226.sdi.lz4.00000000000000000000, size 1712 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000000, size 2490 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/global_variables_193.sdi.lz4.00000000000000000000, size 1448 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/ibdata1.lz4.00000000000000000000, size 56348 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/ibdata1.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/global_variables_193.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/slow_log_226.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/ibdata1.lz4.00000000000000000001, size 25 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/wsrep_streaming_log.ibd.lz4.00000000000000000001, size 47 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/global_variables_193.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/slow_log_226.sdi.lz4.00000000000000000001, size 40 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_140.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/data_lock_waits_170.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_95.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_140.sdi.lz4.00000000000000000000, size 2702 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_95.sdi.lz4.00000000000000000000, size 2202 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/data_lock_waits_170.sdi.lz4.00000000000000000000, size 1924 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000000, size 1777 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_95.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/data_lock_waits_170.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_140.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_158.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_his_95.sdi.lz4.00000000000000000001, size 60 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/data_lock_waits_170.sdi.lz4.00000000000000000001, size 60 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_140.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/keyring_keys_161.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/general_log.CSM.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/threads_119.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/cond_instances_91.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/keyring_keys_161.sdi.lz4.00000000000000000000, size 1275 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/general_log.CSM.lz4.00000000000000000000, size 99 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/threads_119.sdi.lz4.00000000000000000000, size 2719 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/cond_instances_91.sdi.lz4.00000000000000000000, size 1459 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/cond_instances_91.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/threads_119.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/general_log.CSM.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/keyring_keys_161.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/cond_instances_91.sdi.lz4.00000000000000000001, size 58 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/general_log.CSM.lz4.00000000000000000001, size 39 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/threads_119.sdi.lz4.00000000000000000001, size 52 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/keyring_keys_161.sdi.lz4.00000000000000000001, size 57 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/error_log_92.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/user_variables_b_185.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/ib_buffer_pool.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/error_log_92.sdi.lz4.00000000000000000000, size 1988 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/user_variables_b_185.sdi.lz4.00000000000000000000, size 1591 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/ib_buffer_pool.lz4.00000000000000000000, size 1683 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000000, size 2102 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/ib_buffer_pool.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/user_variables_b_185.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/error_log_92.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_133.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/user_variables_b_185.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/ib_buffer_pool.lz4.00000000000000000001, size 32 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/error_log_92.sdi.lz4.00000000000000000001, size 53 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/rwlock_instances_110.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/accounts_154.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_status_191.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000000, size 2789 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/rwlock_instances_110.sdi.lz4.00000000000000000000, size 1641 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_status_191.sdi.lz4.00000000000000000000, size 1438 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/accounts_154.sdi.lz4.00000000000000000000, size 1559 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_status_191.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/accounts_154.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/rwlock_instances_110.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_status_191.sdi.lz4.00000000000000000001, size 59 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster_members.ibd.lz4.00000000000000000001, size 49 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/accounts_154.sdi.lz4.00000000000000000001, size 53 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/rwlock_instances_110.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_threads_115.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/data_locks_169.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_174.sdi.lz4.00000000000000000000. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000000, size 2323 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_174.sdi.lz4.00000000000000000000, size 2391 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/data_locks_169.sdi.lz4.00000000000000000000, size 2028 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_threads_115.sdi.lz4.00000000000000000000, size 1740 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_174.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/data_locks_169.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_threads_115.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000001. 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_174.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_threads_115.sdi.lz4.00000000000000000001, size 58 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_137.sdi.lz4.00000000000000000001, size 61 260429 16:56:04 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/data_locks_169.sdi.lz4.00000000000000000001, size 55 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_179.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_cur_93.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000000, size 1703 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000000, size 2543 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_179.sdi.lz4.00000000000000000000, size 1740 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_cur_93.sdi.lz4.00000000000000000000, size 2269 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_cur_93.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_179.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/wsrep_cluster.ibd.lz4.00000000000000000001, size 41 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_cur_93.sdi.lz4.00000000000000000001, size 60 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_179.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_errors_su_149.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_variable_194.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000000, size 1607 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000000, size 1592 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000000, size 1554 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_variable_194.sdi.lz4.00000000000000000000, size 1445 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_variable_194.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_variable_194.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_96.sdi.lz4.00000000000000000001, size 60 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_98.sdi.lz4.00000000000000000001, size 60 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_97.sdi.lz4.00000000000000000001, size 60 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/global_status_190.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/host_cache_105.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/prepared_stateme_184.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/global_status_190.sdi.lz4.00000000000000000000, size 1441 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000000, size 1516 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/prepared_stateme_184.sdi.lz4.00000000000000000000, size 2789 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/host_cache_105.sdi.lz4.00000000000000000000, size 2301 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/prepared_stateme_184.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/host_cache_105.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/global_status_190.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/prepared_stateme_184.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/host_cache_105.sdi.lz4.00000000000000000001, size 55 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_127.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/global_status_190.sdi.lz4.00000000000000000001, size 58 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/hosts_155.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/log_status_183.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/myApp/myApp.ibd.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/hosts_155.sdi.lz4.00000000000000000000, size 1491 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000000, size 2036 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/log_status_183.sdi.lz4.00000000000000000000, size 1289 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/myApp/myApp.ibd.lz4.00000000000000000000, size 2327 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/myApp/myApp.ibd.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/log_status_183.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/hosts_155.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/myApp/myApp.ibd.lz4.00000000000000000001, size 33 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_131.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/log_status_183.sdi.lz4.00000000000000000001, size 55 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/hosts_155.sdi.lz4.00000000000000000001, size 50 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_201.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql.ibd.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_g_162.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_tot_200.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_201.sdi.lz4.00000000000000000000, size 1274 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_g_162.sdi.lz4.00000000000000000000, size 1630 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_tot_200.sdi.lz4.00000000000000000000, size 1254 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql.ibd.lz4.00000000000000000000, size 4545124 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_tot_200.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_g_162.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql.ibd.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_201.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_tot_200.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_g_162.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/malloc_stats_201.sdi.lz4.00000000000000000001, size 57 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql.ibd.lz4.00000000000000000001, size 27 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_163.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_accoun_186.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000000, size 2499 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_accoun_186.sdi.lz4.00000000000000000000, size 1535 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_163.sdi.lz4.00000000000000000000, size 1794 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000000, size 1765 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_accoun_186.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_163.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_transacti_146.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_accoun_186.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_177.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_163.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSV.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000000, size 1729 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSV.lz4.00000000000000000000, size 79 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000000, size 1732 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000000, size 1693 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSV.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/slow_log.CSV.lz4.00000000000000000001, size 36 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_166.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_164.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/memory_summary_b_165.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/mutex_instances_106.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/performance_time_108.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/processlist_109.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/pxc_cluster_view_203.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/mutex_instances_106.sdi.lz4.00000000000000000000, size 1579 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/performance_time_108.sdi.lz4.00000000000000000000, size 1465 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/processlist_109.sdi.lz4.00000000000000000000, size 1978 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/pxc_cluster_view_203.sdi.lz4.00000000000000000000, size 1342 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/pxc_cluster_view_203.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/processlist_109.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/performance_time_108.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/mutex_instances_106.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/pxc_cluster_view_203.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/performance_time_108.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/mutex_instances_106.sdi.lz4.00000000000000000001, size 60 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/processlist_109.sdi.lz4.00000000000000000001, size 56 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/general_log.CSV.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000000, size 1672 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/general_log.CSV.lz4.00000000000000000000, size 82 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000000, size 2125 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000000, size 1553 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/mysql/general_log.CSV.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_138.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/mysql/general_log.CSV.lz4.00000000000000000001, size 39 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_176.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_appl_175.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_173.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_178.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_account__160.sdi.lz4.00000000000000000000. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_173.sdi.lz4.00000000000000000000, size 2549 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_178.sdi.lz4.00000000000000000000, size 1706 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_account__160.sdi.lz4.00000000000000000000, size 1634 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000000, size 1673 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_account__160.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_178.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_173.sdi.lz4.00000000000000000001. 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_account__160.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_178.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_conn_173.sdi.lz4.00000000000000000001, size 61 260429 16:56:05 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_100.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_actors_111.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_123.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_172.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_connect__159.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_actors_111.sdi.lz4.00000000000000000000, size 1565 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_123.sdi.lz4.00000000000000000000, size 1552 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_172.sdi.lz4.00000000000000000000, size 1509 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_connect__159.sdi.lz4.00000000000000000000, size 1629 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_172.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/session_connect__159.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_123.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_actors_111.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/replication_grou_172.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_actors_111.sdi.lz4.00000000000000000001, size 57 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_su_123.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/session_connect__159.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_instrument_113.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_objects_114.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/socket_instances_156.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_157.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_instrument_113.sdi.lz4.00000000000000000000, size 2019 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_157.sdi.lz4.00000000000000000000, size 1874 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_objects_114.sdi.lz4.00000000000000000000, size 1714 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/socket_instances_156.sdi.lz4.00000000000000000000, size 1835 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_157.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/socket_instances_156.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_objects_114.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/setup_instrument_113.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/socket_summary_b_157.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_objects_114.sdi.lz4.00000000000000000001, size 58 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/socket_instances_156.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/setup_instrument_113.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_cu_120.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_user_189.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_thread_188.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_user_189.sdi.lz4.00000000000000000000, size 1472 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_cu_120.sdi.lz4.00000000000000000000, size 1907 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_thread_188.sdi.lz4.00000000000000000000, size 1563 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000000, size 1808 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_thread_188.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/status_by_user_189.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_cu_120.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_thread_188.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/status_by_user_189.sdi.lz4.00000000000000000001, size 59 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_stages_cu_120.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/file_summary_by__103.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/metadata_locks_168.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_handles_167.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/binlog.index.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_handles_167.sdi.lz4.00000000000000000000, size 1735 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000000, size 2908 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/metadata_locks_168.sdi.lz4.00000000000000000000, size 1849 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/binlog.index.lz4.00000000000000000000, size 95 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/binlog.index.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_handles_167.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/metadata_locks_168.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/binlog.index.lz4.00000000000000000001, size 30 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/metadata_locks_168.sdi.lz4.00000000000000000001, size 59 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_handles_167.sdi.lz4.00000000000000000001, size 58 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_statement_130.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_116.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_lock_waits_118.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/tls_channel_stat_199.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_116.sdi.lz4.00000000000000000000, size 1955 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000000, size 1612 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_lock_waits_118.sdi.lz4.00000000000000000000, size 2301 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/tls_channel_stat_199.sdi.lz4.00000000000000000000, size 1299 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_lock_waits_118.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_116.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/tls_channel_stat_199.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_lock_waits_118.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/tls_channel_stat_199.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/table_io_waits_s_116.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/events_waits_sum_99.sdi.lz4.00000000000000000001, size 60 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/user_defined_fun_197.sdi.lz4.00000000000000000000. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/user_defined_fun_197.sdi.lz4.00000000000000000000, size 1546 260429 16:56:06 xbcloud: [0] Downloading pitr-2026-04-29-16:41:47-full/performance_schema/user_defined_fun_197.sdi.lz4.00000000000000000001. 260429 16:56:06 xbcloud: [0] Download successfull pitr-2026-04-29-16:41:47-full/performance_schema/user_defined_fun_197.sdi.lz4.00000000000000000001, size 61 260429 16:56:06 xbcloud: Download completed. + set +o xtrace vault configuration not found + xtrabackup --use-memory=100MB --prepare --rollback-prepared-trx --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_wgsL 2026-04-29T16:56:06.646641-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=34706702 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2026-04-29T16:56:06.646717-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=100MB --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_wgsL xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2026-04-29T16:56:06.646747-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_wgsL/ 2026-04-29T16:56:06.646813-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2026-04-29T16:56:06.656113-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(30354337) 2026-04-29T16:56:06.656847-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-04-29T16:56:06.656858-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-04-29T16:56:06.656862-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-04-29T16:56:06.656886-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-04-29T16:56:06.656895-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-04-29T16:56:06.656901-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-04-29T16:56:06.657063-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2026-04-29T16:56:06.657233-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-04-29T16:56:06.657245-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-04-29T16:56:06.657251-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-04-29T16:56:06.657260-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-04-29T16:56:06.657265-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-04-29T16:56:06.657271-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-04-29T16:56:06.657282-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2026-04-29T16:56:06.657291-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 104857600 bytes for buffer pool (set by --use-memory parameter) 2026-04-29T16:56:06.657322-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2026-04-29T16:56:06.657339-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2026-04-29T16:56:06.657344-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2026-04-29T16:56:06.657350-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2026-04-29T16:56:06.657504-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2026-04-29T16:56:06.657822-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2026-04-29T16:56:06.657858-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2026-04-29T16:56:06.667274-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2026-04-29T16:56:06.667780-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 128.000000M, instances = 1, chunk size =128.000000M 2026-04-29T16:56:06.674540-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2026-04-29T16:56:06.676357-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2026-04-29T16:56:06.705270-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 30354337 in redo log file ./#innodb_redo/#ib_redo0. 2026-04-29T16:56:06.705301-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 30235617 in the system tablespace does not match the log sequence number 30354337 in the redo log files! 2026-04-29T16:56:06.705308-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2026-04-29T16:56:06.705315-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2026-04-29T16:56:06.705423-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 30354018, whereas checkpoint_lsn = 30354337 and start_lsn = 30353920 2026-04-29T16:56:06.705433-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 30354347 2026-04-29T16:56:06.723313-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-04-29T16:56:06.725334-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2026-04-29T16:56:06.725378-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2026-04-29T16:56:06.725391-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2026-04-29T16:56:06.828657-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-04-29T16:56:06.830118-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=30354347. 2026-04-29T16:56:06.830128-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2026-04-29T16:56:06.830161-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2026-04-29T16:56:06.859584-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2026-04-29T16:56:06.865516-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2026-04-29T16:56:06.870642-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=30354444 2026-04-29T16:56:06.870698-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-04-29T16:56:06.873185-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2026-04-29T16:56:06.874608-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2026-04-29T16:56:06.876363-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2026-04-29T16:56:06.876412-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5200 2026-04-29T16:56:07.054535-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2026-04-29T16:56:07.054553-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 178144 ms. 2026-04-29T16:56:07.054606-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2026-04-29T16:56:07.054657-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2026-04-29T16:56:07.086441-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2026-04-29T16:56:07.086574-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2026-04-29T16:56:07.120112-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2026-04-29T16:56:07.120274-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.35 started; log sequence number 30354454 2026-04-29T16:56:07.121361-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2026-04-29T16:56:07.126987-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00668353 seconds 2026-04-29T16:56:07.147856-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.0208499 seconds 2026-04-29T16:56:08.148266-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 728a337c-43e9-11f1-ae89-223d0a56030a:42 2026-04-29T16:56:08.148303-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2026-04-29T16:56:08.148350-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2026-04-29T16:56:09.148278-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2026-04-29T16:56:09.249157-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-04-29T16:56:09.261215-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 30354454 2026-04-29T16:56:09.264029-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --force-non-empty-directories --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_wgsL 2026-04-29T16:56:09.278249-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2026-04-29T16:56:09.278311-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_wgsL xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2026-04-29T16:56:09.278346-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_wgsL/ 2026-04-29T16:56:09.278810-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2026-04-29T16:56:09.278848-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2026-04-29T16:56:09.278880-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2026-04-29T16:56:09.278909-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2026-04-29T16:56:09.279072-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2026-04-29T16:56:09.279097-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2026-04-29T16:56:09.279334-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000009 to /datadir//binlog.000009 2026-04-29T16:56:09.279358-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000009 to /datadir//binlog.000009 2026-04-29T16:56:09.279432-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2026-04-29T16:56:09.279453-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2026-04-29T16:56:09.279738-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2026-04-29T16:56:09.279778-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2026-04-29T16:56:09.279809-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-04-29T16:56:09.279837-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-04-29T16:56:09.279915-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-04-29T16:56:09.279947-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-04-29T16:56:09.279970-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2026-04-29T16:56:09.279995-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2026-04-29T16:56:09.280013-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2026-04-29T16:56:09.280031-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2026-04-29T16:56:09.280113-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-04-29T16:56:09.280138-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-04-29T16:56:09.280160-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-04-29T16:56:09.280187-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-04-29T16:56:09.280281-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2026-04-29T16:56:09.280314-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2026-04-29T16:56:09.280352-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2026-04-29T16:56:09.280379-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2026-04-29T16:56:09.280407-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-04-29T16:56:09.280459-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-04-29T16:56:09.280499-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-04-29T16:56:09.280527-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-04-29T16:56:09.280569-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2026-04-29T16:56:09.280593-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2026-04-29T16:56:09.280626-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2026-04-29T16:56:09.280658-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2026-04-29T16:56:09.280689-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2026-04-29T16:56:09.280713-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2026-04-29T16:56:09.280739-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2026-04-29T16:56:09.280764-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2026-04-29T16:56:09.280788-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2026-04-29T16:56:09.280810-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2026-04-29T16:56:09.280838-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2026-04-29T16:56:09.280854-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2026-04-29T16:56:09.280882-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2026-04-29T16:56:09.280906-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2026-04-29T16:56:09.280935-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-04-29T16:56:09.280959-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-04-29T16:56:09.280986-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2026-04-29T16:56:09.281009-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2026-04-29T16:56:09.281036-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-04-29T16:56:09.281063-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-04-29T16:56:09.281087-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2026-04-29T16:56:09.281115-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2026-04-29T16:56:09.281144-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2026-04-29T16:56:09.281170-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2026-04-29T16:56:09.281212-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2026-04-29T16:56:09.281242-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2026-04-29T16:56:09.281272-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-04-29T16:56:09.281292-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-04-29T16:56:09.281312-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-04-29T16:56:09.281341-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-04-29T16:56:09.281364-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2026-04-29T16:56:09.281386-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2026-04-29T16:56:09.281415-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2026-04-29T16:56:09.281442-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2026-04-29T16:56:09.281473-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2026-04-29T16:56:09.281505-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2026-04-29T16:56:09.281536-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2026-04-29T16:56:09.281563-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2026-04-29T16:56:09.281589-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2026-04-29T16:56:09.281613-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2026-04-29T16:56:09.281640-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-04-29T16:56:09.281667-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-04-29T16:56:09.281698-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2026-04-29T16:56:09.281724-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2026-04-29T16:56:09.281752-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2026-04-29T16:56:09.281777-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2026-04-29T16:56:09.281800-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-04-29T16:56:09.281818-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-04-29T16:56:09.281843-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-04-29T16:56:09.281867-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-04-29T16:56:09.281897-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2026-04-29T16:56:09.281921-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2026-04-29T16:56:09.281951-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2026-04-29T16:56:09.281974-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2026-04-29T16:56:09.282003-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-04-29T16:56:09.282032-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-04-29T16:56:09.282061-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-04-29T16:56:09.282090-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-04-29T16:56:09.282121-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2026-04-29T16:56:09.282144-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2026-04-29T16:56:09.282162-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2026-04-29T16:56:09.282188-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2026-04-29T16:56:09.282234-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-04-29T16:56:09.282262-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-04-29T16:56:09.282290-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2026-04-29T16:56:09.282319-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2026-04-29T16:56:09.282346-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2026-04-29T16:56:09.282373-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2026-04-29T16:56:09.282399-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-04-29T16:56:09.282426-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-04-29T16:56:09.282455-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2026-04-29T16:56:09.282481-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2026-04-29T16:56:09.282516-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-04-29T16:56:09.282546-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-04-29T16:56:09.282575-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-04-29T16:56:09.282602-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-04-29T16:56:09.282627-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-04-29T16:56:09.282653-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-04-29T16:56:09.282683-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2026-04-29T16:56:09.282709-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2026-04-29T16:56:09.282738-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2026-04-29T16:56:09.282766-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2026-04-29T16:56:09.282796-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-04-29T16:56:09.282817-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-04-29T16:56:09.282841-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-04-29T16:56:09.282868-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-04-29T16:56:09.282900-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-04-29T16:56:09.282930-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-04-29T16:56:09.282953-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2026-04-29T16:56:09.282982-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2026-04-29T16:56:09.283013-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2026-04-29T16:56:09.283034-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2026-04-29T16:56:09.283055-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2026-04-29T16:56:09.283081-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2026-04-29T16:56:09.283112-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-04-29T16:56:09.283139-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-04-29T16:56:09.283168-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-04-29T16:56:09.283194-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-04-29T16:56:09.283242-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2026-04-29T16:56:09.283265-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2026-04-29T16:56:09.283295-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2026-04-29T16:56:09.283324-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2026-04-29T16:56:09.283351-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2026-04-29T16:56:09.283378-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2026-04-29T16:56:09.283408-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2026-04-29T16:56:09.283435-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2026-04-29T16:56:09.283467-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-04-29T16:56:09.283498-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-04-29T16:56:09.283524-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2026-04-29T16:56:09.283552-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2026-04-29T16:56:09.283583-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2026-04-29T16:56:09.283611-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2026-04-29T16:56:09.283643-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-04-29T16:56:09.283671-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-04-29T16:56:09.283702-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2026-04-29T16:56:09.283730-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2026-04-29T16:56:09.283759-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2026-04-29T16:56:09.283786-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2026-04-29T16:56:09.283816-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-04-29T16:56:09.283844-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-04-29T16:56:09.283872-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-04-29T16:56:09.283899-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-04-29T16:56:09.283927-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2026-04-29T16:56:09.283959-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2026-04-29T16:56:09.283985-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2026-04-29T16:56:09.284010-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2026-04-29T16:56:09.284036-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2026-04-29T16:56:09.284062-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2026-04-29T16:56:09.284088-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2026-04-29T16:56:09.284113-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2026-04-29T16:56:09.284138-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2026-04-29T16:56:09.284162-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2026-04-29T16:56:09.284218-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2026-04-29T16:56:09.284249-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2026-04-29T16:56:09.284279-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2026-04-29T16:56:09.284306-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2026-04-29T16:56:09.284333-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-04-29T16:56:09.284357-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-04-29T16:56:09.284382-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2026-04-29T16:56:09.284408-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2026-04-29T16:56:09.284440-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2026-04-29T16:56:09.284471-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2026-04-29T16:56:09.284507-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-04-29T16:56:09.284538-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-04-29T16:56:09.284566-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-04-29T16:56:09.284593-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-04-29T16:56:09.284624-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2026-04-29T16:56:09.284653-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2026-04-29T16:56:09.284686-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-04-29T16:56:09.284713-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-04-29T16:56:09.284745-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2026-04-29T16:56:09.284773-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2026-04-29T16:56:09.284806-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2026-04-29T16:56:09.284835-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2026-04-29T16:56:09.284867-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-04-29T16:56:09.284893-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-04-29T16:56:09.284920-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2026-04-29T16:56:09.284942-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2026-04-29T16:56:09.284973-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2026-04-29T16:56:09.285002-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2026-04-29T16:56:09.285033-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2026-04-29T16:56:09.285063-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2026-04-29T16:56:09.285091-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2026-04-29T16:56:09.285120-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2026-04-29T16:56:09.285147-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2026-04-29T16:56:09.285173-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2026-04-29T16:56:09.285195-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2026-04-29T16:56:09.285254-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2026-04-29T16:56:09.285290-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-04-29T16:56:09.285322-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-04-29T16:56:09.285355-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2026-04-29T16:56:09.285380-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2026-04-29T16:56:09.285408-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-04-29T16:56:09.285438-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-04-29T16:56:09.285469-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2026-04-29T16:56:09.285502-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2026-04-29T16:56:09.285533-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-04-29T16:56:09.285560-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-04-29T16:56:09.285585-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-04-29T16:56:09.285615-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-04-29T16:56:09.285646-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-04-29T16:56:09.285672-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-04-29T16:56:09.285699-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2026-04-29T16:56:09.285726-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2026-04-29T16:56:09.285757-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2026-04-29T16:56:09.285785-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2026-04-29T16:56:09.285817-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-04-29T16:56:09.285847-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-04-29T16:56:09.285876-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2026-04-29T16:56:09.285904-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2026-04-29T16:56:09.285931-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2026-04-29T16:56:09.285961-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2026-04-29T16:56:09.285992-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-04-29T16:56:09.286022-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-04-29T16:56:09.286050-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-04-29T16:56:09.286080-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-04-29T16:56:09.286110-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-04-29T16:56:09.286135-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-04-29T16:56:09.286159-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2026-04-29T16:56:09.286189-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2026-04-29T16:56:09.286233-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-04-29T16:56:09.286264-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-04-29T16:56:09.286297-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-04-29T16:56:09.286322-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-04-29T16:56:09.286352-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-04-29T16:56:09.286381-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-04-29T16:56:09.286409-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2026-04-29T16:56:09.286437-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2026-04-29T16:56:09.286466-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2026-04-29T16:56:09.286496-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2026-04-29T16:56:09.286521-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-04-29T16:56:09.286548-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-04-29T16:56:09.286578-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-04-29T16:56:09.286606-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-04-29T16:56:09.286635-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2026-04-29T16:56:09.286664-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2026-04-29T16:56:09.286692-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2026-04-29T16:56:09.286718-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2026-04-29T16:56:09.286722-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2026-04-29T16:56:09.286735-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2026-04-29T16:56:09.286761-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2026-04-29T16:56:09.286790-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2026-04-29T16:56:09.286870-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-04-29T16:56:09.286901-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-04-29T16:56:09.286927-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-04-29T16:56:09.286952-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-04-29T16:56:09.286979-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2026-04-29T16:56:09.287006-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2026-04-29T16:56:09.287037-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2026-04-29T16:56:09.287061-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2026-04-29T16:56:09.287087-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-04-29T16:56:09.287112-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-04-29T16:56:09.287141-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-04-29T16:56:09.287165-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-04-29T16:56:09.287195-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-04-29T16:56:09.287247-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-04-29T16:56:09.287280-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-04-29T16:56:09.287309-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-04-29T16:56:09.287336-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-04-29T16:56:09.287367-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-04-29T16:56:09.380596-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.wbqMix4LpP Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) + rm /tmp/tmp.89koEXAKbz /tmp/tmp.wbqMix4LpP + return 0 + wait_for_running pitr-proxysql 2 + local name=pitr-proxysql + let last_pod=1 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 1 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ echo pitr-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/pitr-proxysql-0 condition met waiting for pod/pitr-proxysql-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-1 480 + local pod=pitr-proxysql-1 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-proxysql-1 + local container=proxysql + set +o xtrace pod/pitr-proxysql-1 condition met waiting for pod/pitr-proxysql-1 to become Ready.Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ echo pitr-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-1 + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RGB63kYgYT +++ mktemp ++ local LAST_ERR=/tmp/tmp.jYtO0feIXE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RGB63kYgYT ++ cat /tmp/tmp.jYtO0feIXE ++ rm /tmp/tmp.RGB63kYgYT /tmp/tmp.jYtO0feIXE ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.khUKbUiHWJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.kLbN04L7V6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.khUKbUiHWJ ++ cat /tmp/tmp.kLbN04L7V6 ++ rm /tmp/tmp.khUKbUiHWJ /tmp/tmp.kLbN04L7V6 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Ztgl2FLCi4 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Ff2p1EyBIl +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Ztgl2FLCi4 +++++ cat /tmp/tmp.Ff2p1EyBIl +++++ rm /tmp/tmp.Ztgl2FLCi4 /tmp/tmp.Ff2p1EyBIl +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.sndUbbvM0H ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.0P9Y2sq1ls +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.sndUbbvM0H +++++ cat /tmp/tmp.0P9Y2sq1ls +++++ rm /tmp/tmp.sndUbbvM0H /tmp/tmp.0P9Y2sq1ls +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cFpCPkUXo7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jdu5UzzCOS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cFpCPkUXo7 ++ cat /tmp/tmp.Jdu5UzzCOS ++ rm /tmp/tmp.cFpCPkUXo7 /tmp/tmp.Jdu5UzzCOS ++ return 0 + [[ 2 == \2 ]] + echo + desc 'check data after backup' restore-on-pitr-minio-gtid + set +o xtrace ----------------------------------------------------------------------------------- check data after backup restore-on-pitr-minio-gtid ----------------------------------------------------------------------------------- + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VWS5J552zX +++ mktemp ++ local LAST_ERR=/tmp/tmp.QEE3GjD86L ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VWS5J552zX ++ cat /tmp/tmp.QEE3GjD86L ++ rm /tmp/tmp.VWS5J552zX /tmp/tmp.QEE3GjD86L ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.Iojqr3GPLG/select-2.sql + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VKykJUcmoN +++ mktemp ++ local LAST_ERR=/tmp/tmp.1yJE2Jb0qI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VKykJUcmoN ++ cat /tmp/tmp.1yJE2Jb0qI ++ rm /tmp/tmp.VKykJUcmoN /tmp/tmp.1yJE2Jb0qI ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.Iojqr3GPLG/select-2.sql + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KlUWSYjTLG +++ mktemp ++ local LAST_ERR=/tmp/tmp.RoHJTPNgC0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KlUWSYjTLG ++ cat /tmp/tmp.RoHJTPNgC0 ++ rm /tmp/tmp.KlUWSYjTLG /tmp/tmp.RoHJTPNgC0 ++ return 0 + client_pod=pxc-client-67fc4995bb-cwmpz + wait_pod pxc-client-67fc4995bb-cwmpz + local pod=pxc-client-67fc4995bb-cwmpz + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-cwmpz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.Iojqr3GPLG/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.Iojqr3GPLG/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/compare/select-2.sql /tmp/tmp.Iojqr3GPLG/select-2.sql + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/restore-on-pitr-minio-gtid.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.N5ceLr0X0b ++ mktemp + local LAST_ERR=/tmp/tmp.eQATOwqMra + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/restore-on-pitr-minio-gtid.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.N5ceLr0X0b perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio-gtid" deleted from pitr-18261 namespace + cat /tmp/tmp.eQATOwqMra + rm /tmp/tmp.N5ceLr0X0b /tmp/tmp.eQATOwqMra + return 0 ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-18261 ++ local host=pitr-pxc-0.pitr-pxc.pitr-18261 +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.4543Ba2gA5 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.iQgW3OnDPE ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.4543Ba2gA5 ++++ cat /tmp/tmp.iQgW3OnDPE ++++ rm /tmp/tmp.4543Ba2gA5 /tmp/tmp.iQgW3OnDPE ++++ return 0 +++ client_pod=pxc-client-67fc4995bb-cwmpz +++ wait_pod pxc-client-67fc4995bb-cwmpz +++ local pod=pxc-client-67fc4995bb-cwmpz +++ local max_retry=480 +++ local ns= ++++ echo pxc-client-67fc4995bb-cwmpz ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ grep -E '^(pxc|proxysql)$' +++ local container= +++ set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local 'gtid=6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' ++ [[ ! 6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo '6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' + gtid='6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' + log '[CASE 1]: gtid_executed (after restore): 6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:01:29+0000]' '[CASE' '1]:' gtid_executed '(after' 'restore):' '6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' [2026-04-29T17:01:29+0000] [CASE 1]: gtid_executed (after restore): 6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28 + desc '[CASE 1] PiTR with GTID: OK' + set +o xtrace ----------------------------------------------------------------------------------- [CASE 1] PiTR with GTID: OK ----------------------------------------------------------------------------------- + run_backup pitr on-pitr-minio on-pitr-minio-2 + local cluster=pitr + local backup=on-pitr-minio + local name=on-pitr-minio-2 + '[' -z on-pitr-minio-2 ']' + log 'run backup pxc-backup/on-pitr-minio-2' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:01:29+0000]' run backup pxc-backup/on-pitr-minio-2 [2026-04-29T17:01:29+0000] run backup pxc-backup/on-pitr-minio-2 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/on-pitr-minio.yml + kubectl_bin apply -f - + yq eval '.metadata.name="on-pitr-minio-2"' ++ mktemp + local LAST_OUT=/tmp/tmp.TstRyNwmZv ++ mktemp + local LAST_ERR=/tmp/tmp.nQGHxbRrXl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TstRyNwmZv perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio-2 created + cat /tmp/tmp.nQGHxbRrXl + rm /tmp/tmp.TstRyNwmZv /tmp/tmp.nQGHxbRrXl + return 0 + wait_backup on-pitr-minio-2 + local backup=on-pitr-minio-2 + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio-2 to reach Succeeded state.................................................Succeeded + sleep_with_log 65 + local d=65 + log 'sleeping for 65 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:03:22+0000]' sleeping for 65 seconds [2026-04-29T17:03:22+0000] sleeping for 65 seconds + sleep 65 + check_binlog_gap on-pitr-minio-2 + local backup=on-pitr-minio-2 ++ kubectl get pxc-backup on-pitr-minio-2 -o 'jsonpath={.status.conditions}' ++ grep -c 'Binlog with GTID set' + [[ 0 -eq 1 ]] + desc 'no binlog gap after pxc-backup/on-pitr-minio-2: OK' + set +o xtrace ----------------------------------------------------------------------------------- no binlog gap after pxc-backup/on-pitr-minio-2: OK ----------------------------------------------------------------------------------- ++ run_mysql 'SELECT now();' '-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' ++ local 'command=SELECT now();' ++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' +++ get_client_pod +++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.vox86jFUu3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.shLcSYs0fz +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.vox86jFUu3 +++ cat /tmp/tmp.shLcSYs0fz +++ rm /tmp/tmp.vox86jFUu3 /tmp/tmp.shLcSYs0fz +++ return 0 ++ client_pod=pxc-client-67fc4995bb-cwmpz ++ wait_pod pxc-client-67fc4995bb-cwmpz ++ local pod=pxc-client-67fc4995bb-cwmpz ++ local max_retry=480 ++ local ns= +++ echo pxc-client-67fc4995bb-cwmpz +++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' +++ grep -E '^(pxc|proxysql)$' ++ local container= ++ set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ++ set +o xtrace + time_now='2026-04-29 17:04:39' ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-18261 ++ local host=pitr-pxc-0.pitr-pxc.pitr-18261 +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-18261 -uroot -proot_password' +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.WRe7I0TrDB +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.5OdkGPiOvo ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.WRe7I0TrDB ++++ cat /tmp/tmp.5OdkGPiOvo ++++ rm /tmp/tmp.WRe7I0TrDB /tmp/tmp.5OdkGPiOvo ++++ return 0 +++ client_pod=pxc-client-67fc4995bb-cwmpz +++ wait_pod pxc-client-67fc4995bb-cwmpz +++ local pod=pxc-client-67fc4995bb-cwmpz +++ local max_retry=480 +++ local ns= ++++ echo pxc-client-67fc4995bb-cwmpz ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ grep -E '^(pxc|proxysql)$' +++ local container= +++ set +o xtrace pod/pxc-client-67fc4995bb-cwmpz condition met waiting for pod/pxc-client-67fc4995bb-cwmpz to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local 'gtid=6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' ++ [[ ! 6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo '6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' + gtid='6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' + log '[CASE 2]: gtid_executed: 6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:04:50+0000]' '[CASE' '2]:' gtid_executed: '6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28' [2026-04-29T17:04:50+0000] [CASE 2]: gtid_executed: 6d3f7401-43ec-11f1-ad3c-962afed91d34:1,\n728a337c-43e9-11f1-ae89-223d0a56030a:28 + write_data_for_pitr pitr 100503 3 + local cluster=pitr + local start=100503 + local rows=3 ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MX612x1ezt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.lYiYz28wpX +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.MX612x1ezt +++ cat /tmp/tmp.lYiYz28wpX +++ rm /tmp/tmp.MX612x1ezt /tmp/tmp.lYiYz28wpX +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.jQ64uw4Pn3 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XZk2JvXTf0 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.jQ64uw4Pn3 +++ cat /tmp/tmp.XZk2JvXTf0 +++ rm /tmp/tmp.jQ64uw4Pn3 /tmp/tmp.XZk2JvXTf0 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + local sql= + (( i=0 )) + (( i/2026-04-29 17:04:39/g' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-18261~ + /usr/bin/sed -e 's///g' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr/conf/restore-on-pitr-minio-time.yaml + /usr/bin/sed -e 's///g' ++ mktemp + local LAST_OUT=/tmp/tmp.IFCZ3N3DgQ ++ mktemp + local LAST_ERR=/tmp/tmp.TQ4D0fvwUP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.IFCZ3N3DgQ perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-time created + cat /tmp/tmp.TQ4D0fvwUP + rm /tmp/tmp.IFCZ3N3DgQ /tmp/tmp.TQ4D0fvwUP + return 0 + wait_backup_restore restore-on-pitr-minio-time 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-time + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-time to reach Stopping Cluster state 2026-04-29T17:05:18 pxc-restore/restore-on-pitr-minio-time state: Stopping Cluster + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace .Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-1" not found + wait_for_delete pod/pitr-pxc-0 + local res=pod/pitr-pxc-0 + echo -n 'waiting for pod/pitr-pxc-0 to be deleted' waiting for pod/pitr-pxc-0 to be deleted+ set +o xtrace ...Error from server (NotFound): pods "pitr-pxc-0" not found + wait_backup_restore restore-on-pitr-minio-time 'Point-in-time recovering' + local backup_name=restore-on-pitr-minio-time + local 'target_state=Point-in-time recovering' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-time to reach Point-in-time recovering state 2026-04-29T17:05:58 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:01 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:04 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:07 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:10 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:13 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:16 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:19 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:23 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:26 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:28 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:31 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:33 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:36 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:39 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:42 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:44 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:47 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:49 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:52 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:55 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:06:57 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:00 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:02 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:05 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:07 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:09 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:12 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:14 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:16 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:19 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:21 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:23 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:25 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:27 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:30 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:32 pxc-restore/restore-on-pitr-minio-time state: Restoring 2026-04-29T17:07:34 pxc-restore/restore-on-pitr-minio-time state: Point-in-time recovering + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ........................................................................................................................2026-04-29T16:30:14.103Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.32.13-gke.1362000"} 2026-04-29T16:30:14.104Z INFO setup Manager starting up {"gitCommit": "9bec550e38b53ea8a42441eabf05aff8b718c288", "gitBranch": "PR-2433-9bec550e", "buildTime": "2026-04-29T14:21:52Z", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} 2026-04-29T16:30:14.104Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "", "enabled": ""} 2026-04-29T16:30:14.107Z INFO setup Registering Components. 2026-04-29T16:30:14.877Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-04-29T16:30:14.877Z INFO setup Starting the Cmd. 2026-04-29T16:30:14.878Z INFO controller-runtime.metrics Starting metrics server 2026-04-29T16:30:14.878Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-04-29T16:30:14.878Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-04-29T16:30:14.878Z INFO controller-runtime.webhook Starting webhook server 2026-04-29T16:30:14.878Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-04-29T16:30:14.878Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-04-29T16:30:14.879Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-04-29T16:30:14.979Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:30:15.018Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:30:15.018Z DEBUG events percona-xtradb-cluster-operator-7cd7bb76f4-4zmbw_c7f08da7-1f8d-4a32-a013-f7978f50b6d4 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"3baec3c2-80f2-4448-b273-59bcafa728fd","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1777480215010399009"}, "reason": "LeaderElection"} 2026-04-29T16:30:15.019Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-04-29T16:30:15.019Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-04-29T16:30:15.019Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-04-29T16:30:15.019Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-04-29T16:30:15.120Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-04-29T16:30:15.120Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-04-29T16:30:15.120Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-04-29T16:30:15.120Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-04-29T16:30:15.220Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-04-29T16:30:15.220Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-04-29T16:35:00.774Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "version": "1.20.0"} 2026-04-29T16:35:01.070Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f"} 2026-04-29T16:35:04.270Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f"} 2026-04-29T16:35:04.494Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f"} 2026-04-29T16:35:07.595Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "auto-pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:35:07.613Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:35:07.734Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:35:07.771Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:35:07.827Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:35:07.905Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "pitr-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:35:07.985Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:35:08.117Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d577aeae-4a0f-46cc-9e52-63c5fa74a57f", "object": "pitr-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:35:08.977Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "96e99623-60cf-4d55-9461-dc625466be05", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:35:08.999Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "96e99623-60cf-4d55-9461-dc625466be05", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:36:35.483Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc", "user": "operator"} 2026-04-29T16:36:35.522Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc", "user": "monitor"} 2026-04-29T16:36:35.591Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc"} 2026-04-29T16:36:35.641Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc"} 2026-04-29T16:36:35.682Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc", "user": "xtrabackup"} 2026-04-29T16:36:35.741Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc"} 2026-04-29T16:36:35.782Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc", "user": "replication"} 2026-04-29T16:36:35.791Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f656d48-de11-461a-81f5-58578752babc", "err": "get primary pxc pod: not found"} 2026-04-29T16:36:40.913Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3c572441-2259-4391-a9a4-56b63744a750", "err": "get primary pxc pod: not found"} 2026-04-29T16:36:46.048Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6a2b0c63-a7e8-42e1-868e-702ccfd13c1b", "err": "get primary pxc pod: not found"} 2026-04-29T16:39:12.186Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "63a9d869-960a-4331-8901-4789d07d876d", "user": "root"} 2026-04-29T16:39:12.300Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "63a9d869-960a-4331-8901-4789d07d876d", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}"} 2026-04-29T16:39:12.371Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "63a9d869-960a-4331-8901-4789d07d876d", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T16:39:12.421Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "63a9d869-960a-4331-8901-4789d07d876d", "new version": "8.0.43-34.1"} 2026-04-29T16:39:13.080Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "46ab1fe5-8082-4683-8d07-cfa6c7c37675", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}", "hashChanged": false, "metaChanged": true}   &v1.Service{    TypeMeta: v1.TypeMeta{ -  Kind: "Service", +  Kind: "", -  APIVersion: "v1", +  APIVersion: "",    },    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "0694e550-77f3-4694-a536-6c9b606c23ef", +  UID: "", -  ResourceVersion: "1777480752364431024", +  ResourceVersion: "",    Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:39:12 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil,    Annotations: map[string]string{ -  "cloud.google.com/neg": `{"ingress":true}`,    "percona.com/last-config-hash": "eyJwb3J0cyI6W3sibmFtZSI6Imh0dHAiLCJwb3J0Ijo4MDgwLCJ0YXJnZXRQb3J0"...,    },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "v1", -  Time: s"2026-04-29 16:39:12 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.ServiceSpec{    Ports: []v1.ServicePort{    {    Name: "http", -  Protocol: "TCP", +  Protocol: "",    AppProtocol: nil,    Port: 8080,    TargetPort: intstr.IntOrString{    Type: 0, -  IntVal: 8080, +  IntVal: 0,    StrVal: "",    },    NodePort: 0,    },    },    Selector: {"app.kubernetes.io/component": "pitr", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    ClusterIP: "34.118.236.105", -  ClusterIPs: []string{"34.118.236.105"}, +  ClusterIPs: nil,    Type: "ClusterIP",    ExternalIPs: nil, -  SessionAffinity: "None", +  SessionAffinity: "",    LoadBalancerIP: "",    LoadBalancerSourceRanges: nil,    ... // 3 identical fields    PublishNotReadyAddresses: false,    SessionAffinityConfig: nil, -  IPFamilies: []v1.IPFamily{"IPv4"}, +  IPFamilies: nil, -  IPFamilyPolicy: &"SingleStack", +  IPFamilyPolicy: nil,    AllocateLoadBalancerNodePorts: nil,    LoadBalancerClass: nil, -  InternalTrafficPolicy: &"Cluster", +  InternalTrafficPolicy: nil,    TrafficDistribution: nil,    },    Status: {},   } 2026-04-29T16:39:13.888Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "63a9d869-960a-4331-8901-4789d07d876d"} 2026-04-29T16:39:20.835Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "70c09ec7-7904-4605-a5ad-62a21c4a7e8e"} 2026-04-29T16:39:25.862Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b696e40b-a730-4c7b-8713-d397292bedaa"} 2026-04-29T16:39:31.379Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c6915d55-fb64-4e24-903f-4410b58020f0"} 2026-04-29T16:39:36.647Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "334c897a-30e7-4bba-a9dc-f58e74ea28d4"} 2026-04-29T16:39:42.059Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b87e355a-72cb-41dc-8e3e-2e89cb2daf61"} 2026-04-29T16:39:47.334Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e7657f29-719e-4c5b-85dc-f94bce3c9d6f"} 2026-04-29T16:39:52.653Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e300f2a1-b913-4025-9ca1-a931a3765248"} 2026-04-29T16:39:57.954Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "5cefffea-f369-435a-977b-f7110252d0d6"} 2026-04-29T16:40:03.277Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "0869657f-44c1-438f-8175-833c002d8044"} 2026-04-29T16:40:08.255Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "86674d32-6d55-4961-b0f9-6a4b36c1689a"} 2026-04-29T16:40:13.471Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "243dcf30-73b7-40e0-b833-99663519f00f"} 2026-04-29T16:40:19.547Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e6bf0edd-7eaa-45cb-ae4e-10276b13cb7f"} 2026-04-29T16:40:24.074Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "926bb7ca-07b7-4c3c-92ff-7ca4fb489625"} 2026-04-29T16:40:29.464Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "30335bc8-3279-43f2-81c5-111b1c1db353"} 2026-04-29T16:40:34.673Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c2c845ed-de31-4eb7-8724-6b9253aea98f"} 2026-04-29T16:40:39.667Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e7fe0ff-ae70-40d5-a9f4-376d95bb2cb8"} 2026-04-29T16:40:44.885Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ff2e659c-2516-4ad4-bd5e-7d914d83ff6b"} 2026-04-29T16:40:50.482Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "debb00fa-0828-495e-bbab-71e791602e31"} 2026-04-29T16:40:55.831Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d14d3569-7b51-4cc0-8aae-a4439ecc3563"} 2026-04-29T16:41:00.980Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8662a67e-28f3-47d7-aee7-bf5f20adf47a"} 2026-04-29T16:41:06.264Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "30c27d87-42e4-46d9-af70-2052529fa51a"} 2026-04-29T16:41:11.481Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ede8091a-e4d5-4a5e-bcd9-eb330d9464f2"} 2026-04-29T16:41:16.678Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8b11bded-fe0b-4d8c-9a56-f632b1a66f4e"} 2026-04-29T16:41:22.240Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "853dc175-3e01-457b-99d7-7f95c1d92fe0"} 2026-04-29T16:41:27.687Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c83ccde9-f30a-4a5a-980e-48653780ca2d"} 2026-04-29T16:41:32.768Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "343e480b-0284-48a5-87ed-5cdddde1470a"} 2026-04-29T16:41:38.051Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "efb56d93-47a1-4d35-9484-e09318547a88"} 2026-04-29T16:41:43.573Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1417733a-8e82-4ce1-9481-faeaee23e203"} 2026-04-29T16:41:47.612Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "c2be9fd4-8585-41e0-9467-18d17d271939", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:41:47.662Z INFO Created a new backup job {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "c2be9fd4-8585-41e0-9467-18d17d271939", "namespace": "pitr-18261", "name": "xb-on-pitr-minio"} 2026-04-29T16:41:48.873Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ac026b46-e85f-4c5b-8aed-2255d25a0ecd"} 2026-04-29T16:41:52.662Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "0c256824-820d-4ac2-b22b-26551c11eda2", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:41:52.737Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "80a8caff-3102-4b6f-839f-940cf1f8d110", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:41:54.245Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c728c0bd-d48a-4308-8482-3861eadb5d73"} 2026-04-29T16:41:57.777Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "53ccb45c-fa67-497b-bf70-d72d60875160", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:41:59.176Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "85c33436-62fd-4909-86f6-a32a10fb6c5d"} 2026-04-29T16:42:02.827Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "436776c5-70a5-4d2b-99f2-adc7013ee9b4", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:42:04.766Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b713cdd4-efe2-46fa-b245-6c50535b3c9a"} 2026-04-29T16:42:07.873Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "6cad8be2-9747-461b-a229-f78bd1c47cc8", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:42:09.736Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d5cf3425-b69f-442a-b51f-f8897b00a79d"} 2026-04-29T16:42:12.916Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "a9af32b2-fad0-4b7f-954e-5c2f467d6b18", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:42:15.080Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4a069910-bcbd-44bd-8fbf-fa39aaf45f35"} 2026-04-29T16:42:17.954Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "bea59b69-2164-425f-8095-969d2a1d20ea", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:42:18.059Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "83a66d70-c968-473a-9776-2aa87062460e", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:42:20.829Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6ec87c99-8a87-4cf0-9106-f90bd879bafd"} 2026-04-29T16:42:23.100Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "e6dedb15-72be-4377-94ca-92e5b8d5ec92", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:42:23.137Z INFO Backup succeeded {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "e6dedb15-72be-4377-94ca-92e5b8d5ec92", "job": "xb-on-pitr-minio"} 2026-04-29T16:42:23.137Z DEBUG Removing binlog gap file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "e6dedb15-72be-4377-94ca-92e5b8d5ec92", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-7f6d788949-b6fbt"} 2026-04-29T16:42:24.659Z DEBUG Removing binlog timeline file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "e6dedb15-72be-4377-94ca-92e5b8d5ec92", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-7f6d788949-b6fbt"} 2026-04-29T16:42:24.942Z DEBUG Removing mysql-init secret {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio", "reconcileID": "e6dedb15-72be-4377-94ca-92e5b8d5ec92", "job": "xb-on-pitr-minio", "secret": "pitr-mysql-init"} 2026-04-29T16:42:26.074Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "41b377f1-05e0-483b-a97e-cb614019f471"} 2026-04-29T16:42:31.441Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9a81e97a-6b5f-4ae1-98d1-f534c4bd03c3"} 2026-04-29T16:42:37.080Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c40afb49-27f0-44a3-92e8-bf5051e51b95"} 2026-04-29T16:42:42.988Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1551d8bc-7804-42a7-a65f-f72665fb06a0"} 2026-04-29T16:42:49.061Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6b081e80-1e35-401e-b22b-7c8f20bd44e9"} 2026-04-29T16:42:54.464Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1cf54f30-03a1-449c-9147-eabb8cc3a63b"} 2026-04-29T16:43:00.088Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "51c36572-ce50-4a2a-a3a3-22f2f9c93bfb"} 2026-04-29T16:43:06.183Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "0edafde9-bab3-4780-be42-21022666b11d"} 2026-04-29T16:43:10.877Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "212491a9-e31e-45c2-89b4-4760b54b9700", "latest": "2026-04-29 16:43:05 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:43:12.085Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "212491a9-e31e-45c2-89b4-4760b54b9700"} 2026-04-29T16:43:17.932Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a3f6578c-511f-4f0d-bb99-51637905b72f"} 2026-04-29T16:43:23.678Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "107e6bb2-1d5e-40f3-a445-019981a6a020"} 2026-04-29T16:43:29.667Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e436381c-37cd-41f5-9e89-6dce1784a77b"} 2026-04-29T16:43:35.650Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "eaec476d-79b8-4f70-aae4-c85cd5b67b8f"} 2026-04-29T16:43:41.734Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "5c098049-324d-4be6-9986-41f2c07bd606"} 2026-04-29T16:43:47.636Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "94d3270c-b4bd-4026-abdd-4d4f49616e0c"} 2026-04-29T16:43:52.786Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3621a4c7-c1ca-490b-92c8-99960174d33f"} 2026-04-29T16:43:58.684Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c0bc7d0d-4091-4c02-8f30-f150fd19c8c5"} 2026-04-29T16:44:04.660Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f131a5cb-50cf-482c-9ecc-22f6b631cd7b"} 2026-04-29T16:44:09.775Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "502bfede-ccd4-4582-96f4-bd108fc2e0d2"} 2026-04-29T16:44:14.769Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fac70b4d-07b9-4918-bd96-be33b15470ea", "user": "xtrabackup"} 2026-04-29T16:44:14.796Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fac70b4d-07b9-4918-bd96-be33b15470ea", "user": "xtrabackup"} 2026-04-29T16:44:14.818Z INFO MySQL init secret created {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fac70b4d-07b9-4918-bd96-be33b15470ea", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T16:44:14.846Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fac70b4d-07b9-4918-bd96-be33b15470ea", "user": "xtrabackup"} 2026-04-29T16:44:14.867Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fac70b4d-07b9-4918-bd96-be33b15470ea", "user": "xtrabackup"} 2026-04-29T16:44:14.873Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fac70b4d-07b9-4918-bd96-be33b15470ea", "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258"} 2026-04-29T16:44:14.876Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fac70b4d-07b9-4918-bd96-be33b15470ea", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777480748599023009", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJiN2Q0YjZhNWZiMmVhYTFmOTJjZWI2ZGEzNzVlNTFiMSIsInBlcmNvbmEuY29tL3NzbC1pbnRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:35:07 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:39:08 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc", +  "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258",    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1",    "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-57d58bbb78", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-57d58bbb78", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:44:15.990Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "06a5d80a-b6de-4292-b5a2-ac5e78f748f5"} 2026-04-29T16:45:20.494Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6037d84b-ed92-487f-8797-0792d2685c44", "latest": "2026-04-29 16:45:18 +0000 UTC", "lastBackup": "on-pitr-minio"} [mysql] 2026/04/29 16:45:27 packets.go:58 read tcp 10.159.233.54:38908->10.159.233.60:33062: read: connection reset by peer 2026-04-29T16:45:27.333Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9dac380a-c35f-4713-9d3a-3aefd73c059a", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: invalid connection"} 2026-04-29T16:46:14.772Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9417684e-b5dc-406f-89af-9b567cb356dd", "latest": "2026-04-29 16:45:29 +0000 UTC", "lastBackup": "on-pitr-minio"} [mysql] 2026/04/29 16:46:20 packets.go:58 read tcp 10.159.233.54:49882->10.159.232.41:33062: read: connection reset by peer 2026-04-29T16:46:20.676Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "dcb47192-9254-49c3-95e7-81b0a960fb26", "err": "failed to connect to pod pitr-pxc-0: invalid connection"} 2026-04-29T16:46:26.174Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c460cc6d-fd2f-4834-9555-111256fdf0be", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:46:31.861Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "19174eec-ce78-434d-a9dd-b2766977543e", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:46:37.354Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ac7e6c69-75fd-487c-8bae-654b88438df4", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:46:42.814Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6f05c897-8ec6-404b-af61-f25411cd524a", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:46:48.296Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "128af275-33ab-4dbf-a506-5f2924a7b387", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:46:54.161Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "63b9f4b7-f1d2-4f9a-80f7-c8264a069b7f", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:46:59.641Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3a4543e1-27e9-4d16-ba04-35ad36b9d6a5", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:47:09.017Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a5e1436b-9d3c-4e44-819d-25abbb8c475f"} 2026-04-29T16:47:13.198Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e3a1fb5d-769c-4d3e-9b3d-eb64d2e49b1c", "latest": "2026-04-29 16:45:18 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:47:14.164Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e3a1fb5d-769c-4d3e-9b3d-eb64d2e49b1c"} 2026-04-29T16:47:20.876Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "90ad4137-6f25-4935-8fb6-17cc90378e0c"} 2026-04-29T16:47:23.146Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0", "user": "xtrabackup"} 2026-04-29T16:47:23.173Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0", "user": "xtrabackup"} 2026-04-29T16:47:23.192Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T16:47:23.215Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0", "user": "xtrabackup"} 2026-04-29T16:47:23.239Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0", "user": "xtrabackup"} 2026-04-29T16:47:23.248Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0", "last-applied-secret": "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d"} 2026-04-29T16:47:23.253Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777481221475007009", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:44:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:47:01 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258", +  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1",    "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-74c945f9f7", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-74c945f9f7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:47:25.489Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6874df29-5f3c-4bc8-9059-dc80058a53f0"} 2026-04-29T16:48:13.963Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "0fb27c94-0788-4195-9146-0cd5fc7a0c82", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:48:16.666Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "54ac4e01-853a-479e-9fce-2394c19c975e", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:48:19.903Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e52e201f-ee15-4fb2-935b-b9b6d9435baa", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:48:22.209Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ba88dba8-caba-42b2-abc0-7b182222d5f7", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:48:24.002Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b4308dfe-6b6d-4b4b-8b49-ae7262c9954c", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:48:31.830Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6be5e9d4-de87-41ae-a8ab-48b235362661", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:48:32.222Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6be5e9d4-de87-41ae-a8ab-48b235362661", "latest": "2026-04-29 16:48:26 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:49:24.770Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9eb34edd-32d8-4b04-8c4e-d066110d906d", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:49:25.989Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d313dfd5-e187-4369-b278-2db89859014d", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:49:31.799Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8a09aa55-5b08-43e9-98b0-a72e4d19ebdf", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:49:37.536Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4f36fe23-5ece-43a1-88d2-6cea67821b89", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:49:43.012Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7788f408-a0b9-4a4c-a3a2-ee64c9c4c8e0", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:49:43.727Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "575f4b1c-e75f-4dd2-8351-d8a1691a042d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:49:49.241Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "86eb6f0d-2a07-4341-9cc2-46ba4e40d4d7", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:49:54.962Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3b1dbdbf-b932-4bff-b05b-54f3b5003196", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:50:00.459Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "afebe732-952f-4ae8-9cda-3f1be6a50d75", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:50:01.274Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f725c679-1c5c-4771-b5ac-206e3689ad47", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:50:06.980Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d944726f-6e17-4e06-955c-c390a6681bd6", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:50:07.782Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "baa12e43-f3fb-42a6-830e-141177c0988d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:50:08.495Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a45fd6f6-c12f-498f-b086-0bf8b3dbad57", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:50:13.955Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "45457eba-23f2-4884-8b68-4438e20a4bc3", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:50:23.533Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "763627cd-1163-4031-9312-df7d3be04c70"} 2026-04-29T16:50:29.090Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4a786c10-8946-4fe1-80fd-ec05052d0fb5"} 2026-04-29T16:50:29.963Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be", "user": "xtrabackup"} 2026-04-29T16:50:29.990Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be", "user": "xtrabackup"} 2026-04-29T16:50:30.006Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T16:50:30.033Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be", "user": "xtrabackup"} 2026-04-29T16:50:30.058Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be", "user": "xtrabackup"} 2026-04-29T16:50:30.065Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27"} 2026-04-29T16:50:30.068Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777481417721247009", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiJkZjhjYmI5YmQ2NzM1MmMxMDIyOGY0ODcwNjkwZmU4MDIwYzE5MzNjNjkxMWY3Yjk0NmM0N2UyYjc2MDgwYjI3IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:47:23 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:50:17 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d", +  "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1",    "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-5cb49b4c95", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-5cb49b4c95", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:50:32.583Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3e525679-ff32-4163-805e-24993409c3be"} 2026-04-29T16:51:19.882Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "67da9d1e-4872-4f61-98d5-769976784862", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:21.923Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "931bc6ff-c45f-49fe-9f79-50df23cf8693", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:24.400Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "04d6c9ab-bbfc-4adc-98c4-03918d6fafe5", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:26.673Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1899c0cf-6e0a-4366-bd7e-aebb9e7b1f3e", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:28.879Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4347fe51-af92-4857-be12-25e79d4c4a92", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:31.225Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6c1764fe-5f09-47af-bcd7-7d85b7022c5b", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:33.204Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7f7e13f8-e463-46f2-9183-f43199aef564", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:36.130Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a27bf49d-8fc3-4d65-a6ae-20460d3cbc7e", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:38.863Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4967bc51-eb9a-468f-8c7b-b017301f71b3", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:42.668Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1182dded-0d4e-45c2-9262-511a3c9a6598", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:51:46.672Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "179e78f7-3590-425e-aba7-e1dbd29a750b", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:52:10.326Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e5d61cd8-e738-411b-b136-dbab012fa865", "latest": "2026-04-29 16:52:07 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:52:44.529Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "31039349-c459-4f72-ad71-c775c0b8d28a", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:52:50.275Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "5fbe84f8-8bcd-474c-8fab-134dfb6e3fb7", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:52:55.779Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4d62b1b3-1c99-47d2-a01c-81cf9285aa27", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:53:01.249Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9ec2c538-f3e6-47f9-9f5d-72f61a24482d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:53:06.990Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e1e80259-3a1b-42a3-a8ea-3d7099162f39", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:53:12.690Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "64feab1c-0af9-40e9-a511-2b406b165f82", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:53:18.569Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c401aafc-6c56-4704-9fad-aab8b1e6da32", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:53:24.297Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9ceab1a8-10b1-4366-a68b-a0253f09d266", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:53:29.787Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d09a8dd9-7fc4-4b94-98ca-4d8c5d84feec", "primary name": "pitr-pxc-0.pitr-pxc.pitr-18261.svc.cluster.local"} 2026-04-29T16:53:33.680Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "04aed38f-5048-44b3-92e2-95e207c75617"} 2026-04-29T16:53:38.371Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f46bebe2-a099-4355-a8d9-5735823f615f"} 2026-04-29T16:53:43.987Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "526f0483-481f-43af-a8a4-82d5743b8c23"} 2026-04-29T16:53:50.132Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "13a7443e-4282-4a41-af53-54d3a1e1e8eb"} 2026-04-29T16:53:55.687Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8258c965-ee9f-4228-a09a-0c2df45c59fa"} 2026-04-29T16:54:01.362Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "684bcf7f-3b67-45f3-ab57-ff2c7c28fc6d"} 2026-04-29T16:54:06.036Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d4cc3b6f-1924-42fb-832e-ac7bcb9a2c47", "latest": "2026-04-29 16:54:00 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:54:06.986Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d4cc3b6f-1924-42fb-832e-ac7bcb9a2c47"} 2026-04-29T16:54:12.684Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "40ca9f43-e20e-47ab-9553-b1680b60d59b"} 2026-04-29T16:54:19.155Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f9c8ab60-e183-434f-88ff-1f27d6b3c1f6"} 2026-04-29T16:54:24.430Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8a1fde81-16c8-4622-abc0-6e0a0b362943"} 2026-04-29T16:54:30.258Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ecc95cc9-f45e-495a-96df-947833e9299c"} 2026-04-29T16:54:35.745Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8b4fe9bb-a201-4334-943a-ca47500e80ef"} 2026-04-29T16:54:41.441Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "cd3d8e9c-e42f-4328-96e2-1d90e273e2d6"} 2026-04-29T16:54:47.538Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b4eba1ce-1dd2-48e9-92d6-08a86d526e3b"} 2026-04-29T16:54:53.213Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e917c603-46b7-4cc9-a917-7cec83aab513", "error": "syncusers: ERROR (line:1301) : Could not find any nodes belonging to the cluster with writer hostgroup:11\nERROR (line:572) : Could not find a primary cluster node\n", "errorVerbose": "syncusers: ERROR (line:1301) : Could not find any nodes belonging to the cluster with writer hostgroup:11\nERROR (line:572) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:997\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:834 2026-04-29T16:54:59.137Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "91366f82-388f-4d44-ab51-4bb0911a5363"} 2026-04-29T16:55:05.655Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4feb0e8d-a0ef-42af-8807-13274bcd852f"} 2026-04-29T16:55:11.191Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "acb9a4f5-0619-4fc0-afa0-c21dbbc1a8f9"} 2026-04-29T16:55:11.780Z INFO stopping cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d2442e2b-bf6f-4581-91ed-ec19a76e2917", "cluster": "pitr"} 2026-04-29T16:55:11.904Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e4432d60-47d0-491e-b66b-e1f6a2cf70d4"} 2026-04-29T16:55:11.905Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e4432d60-47d0-491e-b66b-e1f6a2cf70d4", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777481610029919009", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:50:30 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:53:30 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &3, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-bb9f59fc5", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-bb9f59fc5", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:55:11.975Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e4432d60-47d0-491e-b66b-e1f6a2cf70d4", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481407760447017", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:35:07 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:50:07 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-7fdd899fc7", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-7fdd899fc7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T16:55:12.054Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e4432d60-47d0-491e-b66b-e1f6a2cf70d4", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481407760447017", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:35:07 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:50:07 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-7fdd899fc7", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-7fdd899fc7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T16:55:14.623Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e4432d60-47d0-491e-b66b-e1f6a2cf70d4", "error": "exec binlog collector pod pitr-pitr-7f6d788949-b6fbt: failed to execute command in pod: pods \"pitr-pitr-7f6d788949-b6fbt\" not found", "errorVerbose": "pods \"pitr-pitr-7f6d788949-b6fbt\" not found\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.CheckPITRErrors\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:73\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:459\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-7f6d788949-b6fbt\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.CheckPITRErrors\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:78\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:459\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T16:55:53.926Z INFO starting restore {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7b62395f-be6e-41fb-87f6-c5d68d27fa2c", "cluster": "pitr", "backup": "on-pitr-minio"} 2026-04-29T16:55:53.990Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4881ce30-469c-4450-bfb2-ef85f6cd8652", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T16:55:59.008Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c507d587-d3d2-461e-bb7e-082aa4731856", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T16:56:04.022Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8d638780-9411-4bf2-9e1e-c37dd497d4af", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T16:56:09.047Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "43570412-685c-4d43-93d6-6fae29baaa3d", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T16:56:14.066Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b34d9307-0584-4f51-9672-dd5ef2d188ad", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:14.066Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b34d9307-0584-4f51-9672-dd5ef2d188ad", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:14.067Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b34d9307-0584-4f51-9672-dd5ef2d188ad", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:14.261Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "54bd6ea3-ff51-4d53-b267-4f8886f9f490", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777481752283775009", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:55:11 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:55:52 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-bb9f59fc5", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-bb9f59fc5", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:56:14.310Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "54bd6ea3-ff51-4d53-b267-4f8886f9f490", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481714215471017", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJiN2Q0YjZhNWZiMmVhYTFmOTJjZWI2ZGEzNzVlNTFiMSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZTMxYzYzNzdhMDhhNGVmZjVmOTRjMmY3ZWRkOTQxYzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJiN2Q0YjZhNWZiMmVhYTFmOTJjZWI2ZGEzNzVlNTFiMSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZTMxYzYzNzdhMDhhNGVmZjVmOTRjMmY3ZWRkOTQxYzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:55:11 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:55:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-7fdd899fc7", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-7fdd899fc7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:56:14.384Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "54bd6ea3-ff51-4d53-b267-4f8886f9f490", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481714215471017", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJiN2Q0YjZhNWZiMmVhYTFmOTJjZWI2ZGEzNzVlNTFiMSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZTMxYzYzNzdhMDhhNGVmZjVmOTRjMmY3ZWRkOTQxYzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJiN2Q0YjZhNWZiMmVhYTFmOTJjZWI2ZGEzNzVlNTFiMSIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiZTMxYzYzNzdhMDhhNGVmZjVmOTRjMmY3ZWRkOTQxYzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:55:11 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:55:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-7fdd899fc7", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-7fdd899fc7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:56:19.141Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3b8e51d2-3df7-4b17-b633-b7bb9cc4ef08", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:19.141Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3b8e51d2-3df7-4b17-b633-b7bb9cc4ef08", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:19.141Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3b8e51d2-3df7-4b17-b633-b7bb9cc4ef08", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:19.141Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3b8e51d2-3df7-4b17-b633-b7bb9cc4ef08", "cluster": "pitr"} 2026-04-29T16:56:24.159Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fa0ce66c-1bb6-4cf6-90cb-e683dff66896", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:24.159Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fa0ce66c-1bb6-4cf6-90cb-e683dff66896", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:24.160Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fa0ce66c-1bb6-4cf6-90cb-e683dff66896", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:24.160Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fa0ce66c-1bb6-4cf6-90cb-e683dff66896", "cluster": "pitr"} 2026-04-29T16:56:29.179Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ef83dd03-63a1-4f30-8ef6-4b642cb0c27d", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:29.179Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ef83dd03-63a1-4f30-8ef6-4b642cb0c27d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:29.180Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ef83dd03-63a1-4f30-8ef6-4b642cb0c27d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:29.180Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ef83dd03-63a1-4f30-8ef6-4b642cb0c27d", "cluster": "pitr"} 2026-04-29T16:56:34.206Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9e1764ed-e64f-4a3e-ab77-0d4799ea2b8d", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:34.206Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9e1764ed-e64f-4a3e-ab77-0d4799ea2b8d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:34.207Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9e1764ed-e64f-4a3e-ab77-0d4799ea2b8d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:34.207Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9e1764ed-e64f-4a3e-ab77-0d4799ea2b8d", "cluster": "pitr"} 2026-04-29T16:56:39.236Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "717aa5a8-ee7d-4a81-a4f5-311848c03dde", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:39.236Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "717aa5a8-ee7d-4a81-a4f5-311848c03dde", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:39.237Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "717aa5a8-ee7d-4a81-a4f5-311848c03dde", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:39.237Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "717aa5a8-ee7d-4a81-a4f5-311848c03dde", "cluster": "pitr"} 2026-04-29T16:56:44.266Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a338d283-670e-410d-84d0-aaf88966b649", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:44.266Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a338d283-670e-410d-84d0-aaf88966b649", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:44.267Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a338d283-670e-410d-84d0-aaf88966b649", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:44.267Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a338d283-670e-410d-84d0-aaf88966b649", "cluster": "pitr"} 2026-04-29T16:56:49.285Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "57c872fb-cef7-4edf-8bbf-2d3927514051", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:49.285Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "57c872fb-cef7-4edf-8bbf-2d3927514051", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:49.285Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "57c872fb-cef7-4edf-8bbf-2d3927514051", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:49.285Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "57c872fb-cef7-4edf-8bbf-2d3927514051", "cluster": "pitr"} 2026-04-29T16:56:54.322Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "431419c4-16e3-4404-b753-b4b9cb6a2c7d", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:54.322Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "431419c4-16e3-4404-b753-b4b9cb6a2c7d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:54.323Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "431419c4-16e3-4404-b753-b4b9cb6a2c7d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:54.323Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "431419c4-16e3-4404-b753-b4b9cb6a2c7d", "cluster": "pitr"} 2026-04-29T16:56:59.341Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ebec8bdb-7fb8-48cf-b4cb-804ea4792a79", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:56:59.341Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ebec8bdb-7fb8-48cf-b4cb-804ea4792a79", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:59.341Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ebec8bdb-7fb8-48cf-b4cb-804ea4792a79", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:56:59.341Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ebec8bdb-7fb8-48cf-b4cb-804ea4792a79", "cluster": "pitr"} 2026-04-29T16:57:04.361Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f513c0cf-26d8-470d-a220-9e119166e1f7", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:57:04.361Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f513c0cf-26d8-470d-a220-9e119166e1f7", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:04.361Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f513c0cf-26d8-470d-a220-9e119166e1f7", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:04.361Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f513c0cf-26d8-470d-a220-9e119166e1f7", "cluster": "pitr"} 2026-04-29T16:57:09.385Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2733864a-ba06-4caa-bc56-2b2d13dd3e78", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:57:09.385Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2733864a-ba06-4caa-bc56-2b2d13dd3e78", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:09.386Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2733864a-ba06-4caa-bc56-2b2d13dd3e78", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:09.386Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2733864a-ba06-4caa-bc56-2b2d13dd3e78", "cluster": "pitr"} 2026-04-29T16:57:14.402Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5bce3106-b2a1-48c8-a9c0-6d9436941a34", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:57:14.402Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5bce3106-b2a1-48c8-a9c0-6d9436941a34", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:14.402Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5bce3106-b2a1-48c8-a9c0-6d9436941a34", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:14.402Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5bce3106-b2a1-48c8-a9c0-6d9436941a34", "cluster": "pitr"} 2026-04-29T16:57:16.910Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1f571ed4-08b5-4a63-b346-31930899c239"} 2026-04-29T16:57:16.958Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1f571ed4-08b5-4a63-b346-31930899c239", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:57:19.429Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3af014-a22f-4867-ab97-921089861478", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T16:57:19.429Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3af014-a22f-4867-ab97-921089861478", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:19.430Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3af014-a22f-4867-ab97-921089861478", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T16:57:19.430Z INFO point-in-time recovering {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3af014-a22f-4867-ab97-921089861478", "cluster": "pitr"} 2026-04-29T16:57:19.497Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5aafeec6-9539-4df9-a165-32353f464efc", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T16:57:22.103Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "753eccfe-bf54-4eaa-bdf5-0643aff86708"} 2026-04-29T16:57:22.547Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "753eccfe-bf54-4eaa-bdf5-0643aff86708", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:57:24.518Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "15babbff-21cc-4f38-9e84-d85ef48af28a", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T16:57:27.728Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e5b62e3f-a4d5-4e98-b3c5-54e7e3eeb57e", "user": "xtrabackup"} 2026-04-29T16:57:27.734Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e5b62e3f-a4d5-4e98-b3c5-54e7e3eeb57e"} 2026-04-29T16:57:27.778Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e5b62e3f-a4d5-4e98-b3c5-54e7e3eeb57e", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:57:29.536Z INFO starting cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1bd70ae7-7c61-43e4-9063-d708be43cf41", "cluster": "pitr"} 2026-04-29T16:57:29.797Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "83aaeab7-fba8-4d3c-b14b-0f187c37d8f9", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777481831378015009", +  ResourceVersion: "", -  Generation: 6, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:56:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:57:11 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &1, +  Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 6, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0, -  ReadyReplicas: 1, +  ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-bb9f59fc5", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-bb9f59fc5", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 1, +  AvailableReplicas: 0,    },   } 2026-04-29T16:57:29.843Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "83aaeab7-fba8-4d3c-b14b-0f187c37d8f9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481774532703017", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:56:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:56:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-64cc55864c", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-64cc55864c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:57:29.926Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "83aaeab7-fba8-4d3c-b14b-0f187c37d8f9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481774532703017", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:56:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:56:14 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-64cc55864c", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-64cc55864c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T16:57:32.199Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "83aaeab7-fba8-4d3c-b14b-0f187c37d8f9", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:57:34.630Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e387ee78-2a1b-4de5-9234-162affc6c656", "cluster": "pitr"} 2026-04-29T16:57:35.725Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c7cd2211-607a-4742-ae2e-d55b4bb380c9", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T16:57:39.645Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "11975ba1-364b-4a29-a611-360365770f60", "cluster": "pitr"} 2026-04-29T16:57:42.844Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b26c7f16-1102-43df-83a5-13846919b517", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 10.159.234.45:6032: connect: connection refused"} 2026-04-29T16:57:44.661Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "040de185-5df2-4fcf-8754-a88bc5e10cf1", "cluster": "pitr"} 2026-04-29T16:57:49.689Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a8a3e5c2-851d-42db-8420-b5e189d4d95d", "cluster": "pitr"} 2026-04-29T16:57:54.744Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "39f1efa7-a65d-410f-93a8-5636dc2a188e", "cluster": "pitr"} 2026-04-29T16:57:59.767Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "6a6edbee-036e-4901-a148-673a8193e46b", "cluster": "pitr"} 2026-04-29T16:58:04.786Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d4b88a0e-90ae-4668-ac58-9126009472c0", "cluster": "pitr"} 2026-04-29T16:58:09.804Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a1ab72b9-ec90-4298-811f-933b4d76bd34", "cluster": "pitr"} 2026-04-29T16:58:14.823Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "470dbd66-a999-49e1-a688-95d21d37bdbf", "cluster": "pitr"} 2026-04-29T16:58:19.840Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "6234f0eb-69c0-4469-aef3-1466ac3d65e4", "cluster": "pitr"} 2026-04-29T16:58:24.854Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e0c83d9b-73bd-4a3c-980c-36274576b4b9", "cluster": "pitr"} 2026-04-29T16:58:29.870Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4fc29c82-b042-4b2b-993b-6d5788231539", "cluster": "pitr"} 2026-04-29T16:58:34.896Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "aff78652-8554-45a5-a988-d642352cd901", "cluster": "pitr"} 2026-04-29T16:58:39.912Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c1015112-012e-40b4-8e86-17b1a1e74d0e", "cluster": "pitr"} 2026-04-29T16:58:44.928Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2bd7d556-113e-4c41-a3cc-e76eb0453391", "cluster": "pitr"} 2026-04-29T16:58:49.945Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "582110f5-093b-4f35-a72d-2290337ceeab", "cluster": "pitr"} 2026-04-29T16:58:54.967Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "85441027-a39e-4df5-8f38-b563537a46a7", "cluster": "pitr"} 2026-04-29T16:58:59.984Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "eb31dd03-1347-4f35-bf3c-c7caae74d44c", "cluster": "pitr"} 2026-04-29T16:59:05.003Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8a3bf316-cebc-41c6-9aba-cac32eae03b9", "cluster": "pitr"} 2026-04-29T16:59:10.019Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d264ce9e-4b3b-4ecc-b385-c29ead9e2aa6", "cluster": "pitr"} 2026-04-29T16:59:15.034Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f0be63a0-62aa-4239-8a27-87e03561d3af", "cluster": "pitr"} 2026-04-29T16:59:20.059Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d6f63992-6940-4fea-811c-25923b30b37d", "cluster": "pitr"} 2026-04-29T16:59:25.081Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "34aab69a-a921-4458-bd48-9753f3b45756", "cluster": "pitr"} 2026-04-29T16:59:30.099Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "aaf9cf3c-adc7-447a-9868-2e904221fa5f", "cluster": "pitr"} 2026-04-29T16:59:35.120Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "72a304ff-96e3-4fe9-a567-df3d73a6ad2a", "cluster": "pitr"} 2026-04-29T16:59:40.136Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "730e4658-31f9-404c-aac2-c9a0a25ef11d", "cluster": "pitr"} 2026-04-29T16:59:45.179Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9465463f-9b00-4d1d-b717-c6adc988e24d", "cluster": "pitr"} 2026-04-29T16:59:50.200Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "efaaf8ef-daef-4f4d-b182-979c19fec471", "cluster": "pitr"} 2026-04-29T16:59:55.219Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d9da159b-5d7c-4d70-bd53-d0e133fa8c62", "cluster": "pitr"} 2026-04-29T17:00:00.241Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8f42a4ee-7c10-4894-a384-b67b3d00f004", "cluster": "pitr"} 2026-04-29T17:00:05.257Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-gtid", "reconcileID": "094a0ec4-e48e-48e0-8b06-a914ab652d50", "cluster": "pitr"} 2026-04-29T17:00:12.057Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "590776b0-7f5e-46e6-8928-d814576d30d3"} 2026-04-29T17:00:15.613Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9bcf8af6-e877-49b8-a10e-e050ac9290ec", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T17:00:17.267Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9bcf8af6-e877-49b8-a10e-e050ac9290ec"} 2026-04-29T17:00:22.457Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "edf7009a-aba6-4c7c-931a-1f5c28da4002"} 2026-04-29T17:00:27.487Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "121ce3c2-511d-417a-bcfa-29515a8503c5", "latest": "2026-04-29 16:58:14 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:00:28.460Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "121ce3c2-511d-417a-bcfa-29515a8503c5"} 2026-04-29T17:00:34.677Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7b09064d-06b6-4cb0-8783-fc7a24ec3364"} 2026-04-29T17:00:40.974Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7326dd50-83dc-4e83-9977-4c8a0b4e359d"} 2026-04-29T17:00:46.956Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "629d53e5-6f8d-4e49-8723-14d3600f21f6"} 2026-04-29T17:00:52.841Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c97aa80c-54a8-4999-b2c8-f0b5acb15d4f"} 2026-04-29T17:00:58.663Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "07b6af4c-9cca-46a9-98e6-e6ee07bfb58e"} 2026-04-29T17:01:04.763Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "5a9ab695-9cfb-4e6a-a58b-0ae8616f4be2"} 2026-04-29T17:01:10.363Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "20380643-e29f-484d-a003-a8bf3b1226fd"} 2026-04-29T17:01:16.168Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "045ffa6f-ea8f-4ed0-86e6-629b2c509865"} 2026-04-29T17:01:21.664Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8a509578-0b9f-483c-86e8-2a8d9f77ebb0"} 2026-04-29T17:01:27.436Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "012ab010-1562-4dff-ab6b-636a95a0f62f"} 2026-04-29T17:01:30.995Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "8b97342e-11a0-48df-a858-c98c493ebcb1", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:01:31.032Z INFO Created a new backup job {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "8b97342e-11a0-48df-a858-c98c493ebcb1", "namespace": "pitr-18261", "name": "xb-on-pitr-minio-2"} 2026-04-29T17:01:33.226Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f8afb6e1-1a64-4d25-a92a-839bf44b3b83"} 2026-04-29T17:01:36.033Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "753a8674-db2b-4914-9636-5093992f4cd7", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:01:36.099Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "dad0af96-2e8e-48d1-99c7-10adce15ff1c", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:01:39.044Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "40816d6c-7eb8-48f6-9cbd-d6e3473885f0"} 2026-04-29T17:01:41.241Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "1e1e89ef-9e68-460b-992e-c8367b29d36e", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:01:44.935Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a0d6a52f-131d-409a-b170-8df9f27e7fc5"} 2026-04-29T17:01:46.279Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "f5155966-862a-4174-b5c8-dfd10cf2272d", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:01:50.666Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d8800756-e6de-4f42-a5a3-53b4dc51ff1b"} 2026-04-29T17:01:51.320Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "d65b0391-0191-4439-a202-6dbfb54ac7ea", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:01:55.871Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "21a5955f-2227-418c-bcc1-2ec2449540ea"} 2026-04-29T17:01:56.367Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "0c9bfbc4-6b09-4248-9c19-45d3aad49b89", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:01.411Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "3faa294e-f311-4d6a-ade0-4cbd88efaec3", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:01.482Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "d0425fc7-25d8-4653-8709-08f4819267c5", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:01.880Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f5e96b76-8287-4c77-b164-93c0bb2860b2"} 2026-04-29T17:02:06.527Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "b1a0c618-9b80-444e-ade7-11f47016a214", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:06.624Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "e0c829c4-20a4-4337-86c6-c3c81035b215", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:07.866Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ee861704-f863-4c10-b4d7-6b3396448714"} 2026-04-29T17:02:11.677Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "1fd8c3f9-afe8-47e3-bf96-630f20deff15", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:13.681Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b9c6b538-dea1-4f92-b05a-64fc3e8024ff"} 2026-04-29T17:02:16.714Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "0b8be13a-ea07-4c42-8823-2d3b3d5fe270", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:19.674Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6edceb98-8ce6-4850-8c0e-dfb3606eb6bd"} 2026-04-29T17:02:21.818Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "813e0aca-3ff4-4cc0-89fc-5fd676736261", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:25.327Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c775ecd6-c5e3-4f38-93fe-d2079ea7bb24"} 2026-04-29T17:02:26.862Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "b09d83fe-8f30-43ec-bc6c-45521d5dfccb", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:30.638Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e907e0f2-8ffb-4a85-aecd-981e69475f6f"} 2026-04-29T17:02:31.911Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "21dca1af-6315-402d-9196-73582cb3706c", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:36.829Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f1d15391-8533-456a-b5d0-8b3b55a3667f"} 2026-04-29T17:02:36.949Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "f099cd5b-acaf-4f6c-86b7-e9524894a5ec", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:41.962Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "cc2ca2b2-40ab-4fdb-9187-468884a0e0df"} 2026-04-29T17:02:41.986Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "83de02c0-bd17-4a25-a988-b454ccd6466a", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:47.028Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "68782294-527b-43ea-9bcf-21c63d09624c", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:47.863Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "cb8e992e-3afd-4266-ad73-e76c947d9151"} 2026-04-29T17:02:52.064Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "615c53e4-b22e-452c-a258-634de1de3664", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:53.871Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "603b0c0e-bb1b-4a12-ac0d-5c92c8ca0ffd"} 2026-04-29T17:02:57.109Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "92b2e52c-de72-41c2-8cbd-f674911ad291", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:02:59.559Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e6623d08-35bc-469c-872d-d7a80e0a5470"} 2026-04-29T17:03:02.156Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "1f52ffea-f90d-413f-9575-eac4b3811ae0", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:03:05.661Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "07ef360c-f337-49fc-80ff-82e30e690eb7"} 2026-04-29T17:03:07.200Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "40cd981c-62d9-446f-8957-3e501a512beb", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:03:11.269Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "088f11d2-e921-47ab-bbb4-1151c431a1ba"} 2026-04-29T17:03:12.241Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "b3bd5b4c-a740-4437-b576-74f51064b738", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:03:17.144Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8938d05e-7000-4677-9db5-1d5a61d0fb6a"} 2026-04-29T17:03:17.290Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "82978b25-30dc-482f-9f11-9825dee3f9d5", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T17:03:17.330Z INFO Backup succeeded {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "82978b25-30dc-482f-9f11-9825dee3f9d5", "job": "xb-on-pitr-minio-2"} 2026-04-29T17:03:17.330Z DEBUG Removing binlog gap file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "82978b25-30dc-482f-9f11-9825dee3f9d5", "job": "xb-on-pitr-minio-2", "pod": "pitr-pitr-7f6d788949-pdhqj"} 2026-04-29T17:03:19.069Z DEBUG Removing binlog timeline file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "82978b25-30dc-482f-9f11-9825dee3f9d5", "job": "xb-on-pitr-minio-2", "pod": "pitr-pitr-7f6d788949-pdhqj"} 2026-04-29T17:03:19.154Z DEBUG Removing mysql-init secret {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio-2","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "on-pitr-minio-2", "reconcileID": "82978b25-30dc-482f-9f11-9825dee3f9d5", "job": "xb-on-pitr-minio-2", "secret": "pitr-mysql-init"} 2026-04-29T17:03:22.462Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "2c9a8387-7cdb-4fcd-95d3-168d1b3c7306"} 2026-04-29T17:03:28.458Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "5062cbaf-7df0-4e31-8d92-572e38f02370"} 2026-04-29T17:03:34.666Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "61955483-9fb7-49c5-b500-5517eaa42cd2"} 2026-04-29T17:03:39.866Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a567c2b8-e853-4139-b111-1522078e5bbb"} 2026-04-29T17:03:45.867Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7a25e107-50fc-4011-b759-a09d4b6cd745"} 2026-04-29T17:03:51.476Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a272424e-7268-4e2b-a450-c2fa70b90412"} 2026-04-29T17:03:57.470Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c1e3edf4-bbfd-478f-aac2-e5f872d00f42"} 2026-04-29T17:04:03.662Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "5706bfcf-a21c-45f5-b81f-204f2261e48f"} 2026-04-29T17:04:09.081Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "8dd8fcd9-206c-4686-bf21-71730227fdbb"} 2026-04-29T17:04:14.960Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e44fbbf8-d563-4b39-ae94-1c5aa1d98db1"} 2026-04-29T17:04:20.872Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "c12668af-11a4-479d-9c74-c22387af9051"} 2026-04-29T17:04:26.981Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "81c5b33d-ea49-4845-a146-e561f9c0cd01"} 2026-04-29T17:04:32.469Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d0184b0d-c444-447c-a571-392bef571a22"} 2026-04-29T17:04:38.471Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fd585aca-e9ef-4806-bd35-c37f323c85c6"} 2026-04-29T17:04:44.275Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "852fc3f8-70cb-4773-8754-17b8eae01374"} 2026-04-29T17:04:50.161Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d90f1f48-7f58-4abd-846e-73a0e03b9577"} 2026-04-29T17:04:55.861Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "aa53de2e-01d9-4408-b5a7-ffdda078a136"} 2026-04-29T17:05:01.881Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4b626e35-9ad1-44fa-bf35-cc36a626f445"} 2026-04-29T17:05:08.047Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "40412952-1d46-4632-adf9-a0bce5e2de67"} 2026-04-29T17:05:13.960Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a1d9af27-d0a2-499d-b7c2-df4e07c8140f"} 2026-04-29T17:05:16.447Z INFO stopping cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "18c57a2a-79ec-4dd0-91a0-c7cd5cabc1cd", "cluster": "pitr"} 2026-04-29T17:05:16.589Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "250491de-5507-44ef-b6ea-534cbf5bf1d9"} 2026-04-29T17:05:16.590Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "250491de-5507-44ef-b6ea-534cbf5bf1d9", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777482007799407009", +  ResourceVersion: "", -  Generation: 7, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:57:29 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:00:07 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &3, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 7, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-bb9f59fc5", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-bb9f59fc5", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T17:05:16.662Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "250491de-5507-44ef-b6ea-534cbf5bf1d9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481877448607017", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:57:29 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:57:57 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-64cc55864c", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-64cc55864c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T17:05:16.743Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "250491de-5507-44ef-b6ea-534cbf5bf1d9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777481877448607017", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:57:29 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:57:57 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-64cc55864c", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-64cc55864c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T17:05:57.656Z INFO starting restore {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "e146953e-3ae5-4239-95ec-4475feb7ccab", "cluster": "pitr", "backup": "on-pitr-minio-2"} 2026-04-29T17:05:57.722Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "73c9ff98-9f26-48f0-a59e-e90e25081c1a", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:02.744Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "fb39be2f-fe1d-49ca-b895-f048215cbb69", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:07.759Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "4cd4030d-75ba-49d2-9665-9dedf2d45d9b", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:12.778Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "fd988646-0b80-4cca-9dc5-2dc434627580", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:17.795Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "6b9cae32-f4e0-47fe-a952-99549a674d27", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:22.811Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "11fe348f-ab8f-424e-a343-7c83f5142c1f", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:27.825Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "f4aba7c9-c06b-4aa3-96cf-65079a386775", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:32.844Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "d7e5ee39-209b-4548-b855-daffeff06a8d", "job": "restore-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:06:37.868Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "ff92c0c5-f243-4546-a7fe-749eeb6eb1a1", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:37.868Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "ff92c0c5-f243-4546-a7fe-749eeb6eb1a1", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:37.869Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "ff92c0c5-f243-4546-a7fe-749eeb6eb1a1", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:38.065Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "23fabf4a-60cc-46cb-8abb-a3f4a3a93a22", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777482356300015009", +  ResourceVersion: "", -  Generation: 8, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:16 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:56 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 8, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-bb9f59fc5", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-bb9f59fc5", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:06:38.117Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "23fabf4a-60cc-46cb-8abb-a3f4a3a93a22", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777482356300015009", +  ResourceVersion: "", -  Generation: 8, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:16 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:56 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 8, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-bb9f59fc5", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-bb9f59fc5", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:06:42.946Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9496aef0-45ef-43c9-a509-b2116f9e43a1", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:42.946Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9496aef0-45ef-43c9-a509-b2116f9e43a1", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:42.947Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9496aef0-45ef-43c9-a509-b2116f9e43a1", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:42.947Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9496aef0-45ef-43c9-a509-b2116f9e43a1", "cluster": "pitr"} 2026-04-29T17:06:47.968Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "57e06fe3-183f-4ae8-adc8-bce984802892", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:47.968Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "57e06fe3-183f-4ae8-adc8-bce984802892", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:47.969Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "57e06fe3-183f-4ae8-adc8-bce984802892", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:47.969Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "57e06fe3-183f-4ae8-adc8-bce984802892", "cluster": "pitr"} 2026-04-29T17:06:52.985Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "11d0d5ab-738e-49e7-8198-46a297ea2967", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:52.985Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "11d0d5ab-738e-49e7-8198-46a297ea2967", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:52.985Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "11d0d5ab-738e-49e7-8198-46a297ea2967", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:52.985Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "11d0d5ab-738e-49e7-8198-46a297ea2967", "cluster": "pitr"} 2026-04-29T17:06:58.001Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "b6a89f08-73ba-4d97-8f96-dcdfceb60461", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:06:58.001Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "b6a89f08-73ba-4d97-8f96-dcdfceb60461", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:58.001Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "b6a89f08-73ba-4d97-8f96-dcdfceb60461", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:06:58.001Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "b6a89f08-73ba-4d97-8f96-dcdfceb60461", "cluster": "pitr"} 2026-04-29T17:07:03.017Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "fcedf097-a421-4617-8552-a9e8cf1161a6", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:07:03.017Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "fcedf097-a421-4617-8552-a9e8cf1161a6", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:03.018Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "fcedf097-a421-4617-8552-a9e8cf1161a6", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:03.018Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "fcedf097-a421-4617-8552-a9e8cf1161a6", "cluster": "pitr"} 2026-04-29T17:07:08.040Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "6f0ff3e9-1028-49c3-a856-7147eafe1134", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:07:08.040Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "6f0ff3e9-1028-49c3-a856-7147eafe1134", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:08.041Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "6f0ff3e9-1028-49c3-a856-7147eafe1134", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:08.041Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "6f0ff3e9-1028-49c3-a856-7147eafe1134", "cluster": "pitr"} 2026-04-29T17:07:13.057Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "7ca9f146-4c95-409c-b9e2-52fb2468eb42", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:07:13.057Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "7ca9f146-4c95-409c-b9e2-52fb2468eb42", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:13.058Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "7ca9f146-4c95-409c-b9e2-52fb2468eb42", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:13.058Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "7ca9f146-4c95-409c-b9e2-52fb2468eb42", "cluster": "pitr"} 2026-04-29T17:07:18.078Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "80d39a6d-de2e-4fe8-9a96-822777ab34aa", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:07:18.078Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "80d39a6d-de2e-4fe8-9a96-822777ab34aa", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:18.078Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "80d39a6d-de2e-4fe8-9a96-822777ab34aa", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:18.078Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "80d39a6d-de2e-4fe8-9a96-822777ab34aa", "cluster": "pitr"} 2026-04-29T17:07:23.099Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "0aaeb6de-4c79-406b-aa53-f9718aef2cdd", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:07:23.099Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "0aaeb6de-4c79-406b-aa53-f9718aef2cdd", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:23.100Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "0aaeb6de-4c79-406b-aa53-f9718aef2cdd", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:23.100Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "0aaeb6de-4c79-406b-aa53-f9718aef2cdd", "cluster": "pitr"} 2026-04-29T17:07:28.121Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "adf301d6-4a9b-4a09-9f17-154524692794", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:07:28.121Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "adf301d6-4a9b-4a09-9f17-154524692794", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:28.122Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "adf301d6-4a9b-4a09-9f17-154524692794", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:28.122Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "adf301d6-4a9b-4a09-9f17-154524692794", "cluster": "pitr"} 2026-04-29T17:07:30.351Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d2c5ea81-daeb-47cf-a349-27586b89af40"} 2026-04-29T17:07:30.422Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "d2c5ea81-daeb-47cf-a349-27586b89af40", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T17:07:33.148Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9bdf364c-b360-4979-a47a-c14b93221393", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:07:33.148Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9bdf364c-b360-4979-a47a-c14b93221393", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:33.148Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9bdf364c-b360-4979-a47a-c14b93221393", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:07:33.148Z INFO point-in-time recovering {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "9bdf364c-b360-4979-a47a-c14b93221393", "cluster": "pitr"} 2026-04-29T17:07:33.215Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "30c10636-fade-49ea-acdc-2c3002ae670d", "job": "pitr-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:07:35.566Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f11ab9fb-1b9a-441e-9f68-ee8f16571cb5"} 2026-04-29T17:07:35.609Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f11ab9fb-1b9a-441e-9f68-ee8f16571cb5", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T17:07:38.236Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "4e0b57d7-ff41-435e-a082-4e05cf6dba5d", "job": "pitr-job-restore-on-pitr-minio-time-pitr"} 2026-04-29T17:07:40.765Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f70d4fd3-c73e-4cd1-8a80-0ffe49bb8168"} 2026-04-29T17:07:40.817Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f70d4fd3-c73e-4cd1-8a80-0ffe49bb8168", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T17:07:43.258Z INFO starting cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "6f414926-ed5b-4a39-b9ff-d85c5ab6faf4", "cluster": "pitr"} 2026-04-29T17:07:43.506Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fb99617e-a7c2-49d4-9a24-3f61baedf2c6", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "c9484faa-4eac-4040-b845-9f41d5cd0207", +  UID: "", -  ResourceVersion: "1777482448397871009", +  ResourceVersion: "", -  Generation: 9, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:06:38 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:07:28 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &1, +  Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 9, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0, -  ReadyReplicas: 1, +  ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-bb9f59fc5", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-bb9f59fc5", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 1, +  AvailableReplicas: 0,    },   } 2026-04-29T17:07:43.577Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fb99617e-a7c2-49d4-9a24-3f61baedf2c6", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777482319483535017", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:16 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:19 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-64cc55864c", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-64cc55864c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:07:43.637Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fb99617e-a7c2-49d4-9a24-3f61baedf2c6", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-18261",    SelfLink: "", -  UID: "a4ad63e2-ed40-4c8c-a952-c113a0ccbe2e", +  UID: "", -  ResourceVersion: "1777482319483535017", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:35:07 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "404973a9-59b0-4969-8372-1a823c8e3651", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:16 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:19 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "b7d4b6a5fb2eaa1f92ceb6da375e51b1", "percona.com/ssl-internal-hash": "e31c6377a08a4eff5f94c2f7edd941c5"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-64cc55864c", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-64cc55864c", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:07:45.905Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "fb99617e-a7c2-49d4-9a24-3f61baedf2c6", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T17:07:48.342Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "4d9e5de3-db9e-4a83-98d3-f5b0eceaab65", "cluster": "pitr"} 2026-04-29T17:07:49.457Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9dff6f40-a69a-41b5-b96b-71e5d0421f92", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-18261 on 34.118.224.10:53: no such host"} 2026-04-29T17:07:53.364Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "0a5ee0b6-5e30-4c25-b671-03f5acbfd810", "cluster": "pitr"} 2026-04-29T17:07:56.707Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "00258f6a-0343-4cb9-b02b-16483d4cbcda", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 10.159.234.47:6032: connect: connection refused"} 2026-04-29T17:07:58.405Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "279408f8-6d12-4f4f-b690-de4315a22aed", "cluster": "pitr"} 2026-04-29T17:08:03.427Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "63b1298f-835f-49a1-8692-7176fcdbd566", "cluster": "pitr"} 2026-04-29T17:08:08.548Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "cac314a2-0c5e-4663-9ec1-36921e44dd5b", "cluster": "pitr"} 2026-04-29T17:08:13.577Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "6207f6a8-3e6f-4583-9b8d-b495e00fddd6", "cluster": "pitr"} 2026-04-29T17:08:18.594Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "1d7275af-2c3d-43f0-8008-ef02a3c10f93", "cluster": "pitr"} 2026-04-29T17:08:23.730Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "87174093-98cb-4100-9e16-fc64c896ce45", "cluster": "pitr"} 2026-04-29T17:08:28.745Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "8432f26a-2a34-4e6e-b1b3-e5b6ec365a9a", "cluster": "pitr"} 2026-04-29T17:08:33.766Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "16621ca6-c1b0-4ecf-8021-f1c73fc3833f", "cluster": "pitr"} 2026-04-29T17:08:38.812Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "5cd35b67-0485-4c21-bcf3-697d3f334037", "cluster": "pitr"} 2026-04-29T17:08:43.827Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "2b1d92c6-33b8-4cd4-95a1-d9c6e597cf11", "cluster": "pitr"} 2026-04-29T17:08:48.842Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "4e36b98a-cbf1-46a0-8885-112d3f230279", "cluster": "pitr"} 2026-04-29T17:08:53.893Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "078f1913-155a-4f1d-93e2-083eacf7d5e7", "cluster": "pitr"} 2026-04-29T17:08:58.908Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "817f5681-af77-4253-9a8a-a9fbbb40fd21", "cluster": "pitr"} 2026-04-29T17:09:03.924Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "7d6816f2-8d19-461d-8824-22bb5602beca", "cluster": "pitr"} 2026-04-29T17:09:09.002Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "e22f33ff-6b0f-4fe7-8724-046848afc37d", "cluster": "pitr"} 2026-04-29T17:09:14.018Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "08ffd0d0-16ac-45d0-99a0-15bbe0d6b3c0", "cluster": "pitr"} 2026-04-29T17:09:19.040Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "d13b2864-599b-4fd6-990f-7a64f42fb346", "cluster": "pitr"} 2026-04-29T17:09:24.117Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "54e2df0c-d8f7-40b4-b00b-f73152db5467", "cluster": "pitr"} 2026-04-29T17:09:29.135Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "075af4f4-b8ad-4c17-bd87-630b12410c81", "cluster": "pitr"} 2026-04-29T17:09:34.150Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "3a035985-1fc8-4063-8f6a-c2e506eca592", "cluster": "pitr"} 2026-04-29T17:09:39.175Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "88b0ec09-5b20-4306-aadc-2b6fc28b2fac", "cluster": "pitr"} 2026-04-29T17:09:44.193Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "7431e44a-94f0-4aec-bfe5-8c70af82e752", "cluster": "pitr"} 2026-04-29T17:09:49.212Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "b78b44e0-2449-469f-b14f-d1e2a23e810a", "cluster": "pitr"} 2026-04-29T17:09:54.232Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "e4133f83-9255-4ca2-a67c-f064a1fe27c5", "cluster": "pitr"} 2026-04-29T17:09:59.260Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "62f02cb0-43d3-4735-9622-f6c7860fd069", "cluster": "pitr"} 2026-04-29T17:10:04.278Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "3564711a-220d-464b-9e84-1a07c929ac71", "cluster": "pitr"} 2026-04-29T17:10:09.298Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "292cebee-8dfd-44d0-a703-9c2b05536ac8", "cluster": "pitr"} 2026-04-29T17:10:14.312Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "3f5a922f-b4b8-499e-ad2f-b8cec8a39a79", "cluster": "pitr"} 2026-04-29T17:10:19.329Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-time","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "restore-on-pitr-minio-time", "reconcileID": "5da4dc3c-5b68-4ee4-9cd3-96ccebf6f3e3", "cluster": "pitr"} 2026-04-29T17:10:25.093Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1d91db5b-feaa-4f19-b3b0-0f5396a9400b"} 2026-04-29T17:10:28.569Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "44a06232-a75f-4ba4-8a02-c57673b40a14", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T17:10:30.430Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "44a06232-a75f-4ba4-8a02-c57673b40a14"} 2026-04-29T17:10:35.943Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "455d8165-360d-4cfb-a6d9-fc3e55a50c19"} 2026-04-29T17:10:41.439Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "dd4ed4b7-a5c6-4406-9ea9-3542f8d53ac8"} 2026-04-29T17:10:47.519Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7294cd25-b0a0-4df8-b876-c32d84e12e15"} 2026-04-29T17:10:53.128Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "6b1f9bc9-cf22-4e71-a3dd-3a86e0b33fb8"} 2026-04-29T17:10:59.024Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "2d18977c-6fd1-4932-aa66-fb968b069ede"} 2026-04-29T17:11:05.430Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "2d6fb546-795f-456f-b0dd-9d268bfffb8e"} 2026-04-29T17:11:11.049Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "51578e6d-97ed-4493-934f-91e8d5c77cbf"} 2026-04-29T17:11:16.939Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "4a3ae477-fb3c-44b2-9951-0e75ff1d7d90"} 2026-04-29T17:11:22.595Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b236d197-58b9-4be4-b047-273104d1ab0a"} 2026-04-29T17:11:28.736Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "041311db-6615-40af-856e-a98f4cf61b77"} 2026-04-29T17:11:34.537Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "68d4bb23-278e-409f-8cb2-52218c21d53b"} 2026-04-29T17:11:40.898Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "0cba6a0b-71c6-44c7-8691-383b2f028466"} 2026-04-29T17:11:46.537Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "1ab365f8-d2d1-4b74-9815-53199e5dadc9"} 2026-04-29T17:11:52.638Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "ceb5e455-b9f2-44f5-966e-8d522f7ff4f7"} 2026-04-29T17:11:58.349Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "267882ee-fa9c-4197-9ab2-0c2b88edc944"} 2026-04-29T17:12:04.926Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "f2828af4-beba-45e9-836b-aa8ffa121d60"} 2026-04-29T17:12:10.851Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "229c2356-720e-4672-a12a-58b037edaf79"} 2026-04-29T17:12:16.999Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7d180fb9-4f72-485b-8e19-ffee2bb69b4f"} 2026-04-29T17:12:22.837Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "2bb939dd-08f9-4643-aea0-78a32920c64e"} 2026-04-29T17:12:28.700Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e131adf4-699b-478b-95e0-c26df6728dc7"} 2026-04-29T17:12:34.546Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "7984f38a-67b4-46d5-b705-edba9f410587"} 2026-04-29T17:12:39.942Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "62ceca28-1725-4a13-b7eb-c35bf11cb8ef"} 2026-04-29T17:12:45.830Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "b79fc7c9-1818-47fc-ba41-34a6b7b36c33"} 2026-04-29T17:12:51.745Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9345b488-fc9f-4709-9e54-d1e88b205ef0"} 2026-04-29T17:12:57.320Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "e820d4c4-781b-4b5f-8458-cbe954a2596f"} 2026-04-29T17:13:03.710Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "a70cc082-cdb6-4a45-bd6d-01f470587628"} 2026-04-29T17:13:09.146Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "21967f1e-c925-4573-bdd0-61a992e46b2b"} 2026-04-29T17:13:15.137Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "3c73931a-8f9d-4bb6-8c7b-dc3355fdd61d"} 2026-04-29T17:13:21.047Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "dc7d5183-25da-4e1f-83c9-dd352bd4be8b"} 2026-04-29T17:13:27.351Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "2ae9dc75-9b9f-4c93-8e0b-87c4952185a6"} 2026-04-29T17:13:33.340Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "cfd1a3ae-938c-42a4-8c8f-ccc475946bd9"} 2026-04-29T17:13:39.042Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "be8b810e-20b0-48f3-b3dd-79e9df99703f"} 2026-04-29T17:13:45.150Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-18261"}, "namespace": "pitr-18261", "name": "pitr", "reconcileID": "9da39da8-672a-4b55-b0af-9f71a8285096"} max retry count 120 reached. something went wrong with operator or kubernetes cluster