Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/logs/pitr-pxb-8-0.log Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 + PXCO_FEATURE_GATES=XtrabackupSidecar=true + GTID_PATTERN='[A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + main + create_infra pitr-pxb-531 + local ns=pitr-pxb-531 + '[' -n pxc-operator ']' + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + kubectl patch pxc -n pitr-pxb-18464 pitr --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/pitr patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.JEO6lRbVP6 ++ mktemp + local LAST_ERR=/tmp/tmp.RSgw3whphr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JEO6lRbVP6 perconaxtradbcluster.pxc.percona.com "pitr" deleted from pitr-pxb-18464 namespace + cat /tmp/tmp.RSgw3whphr + rm /tmp/tmp.JEO6lRbVP6 /tmp/tmp.RSgw3whphr + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.a67Nu2GPyj ++ mktemp + local LAST_ERR=/tmp/tmp.6u7OEP5itW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.a67Nu2GPyj perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio" deleted from pitr-pxb-18464 namespace perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio-2" deleted from pitr-pxb-18464 namespace perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio-3" deleted from pitr-pxb-18464 namespace + cat /tmp/tmp.6u7OEP5itW + rm /tmp/tmp.a67Nu2GPyj /tmp/tmp.6u7OEP5itW + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.ibrsug6rdZ ++ mktemp + local LAST_ERR=/tmp/tmp.6LGMZBoCS2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ibrsug6rdZ perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio" deleted from pitr-pxb-18464 namespace + cat /tmp/tmp.6LGMZBoCS2 + rm /tmp/tmp.ibrsug6rdZ /tmp/tmp.6LGMZBoCS2 + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ awk '-F ' '{print $2}' ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// ++ tail -n1 + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrolebinding ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + xargs kubectl delete ns ++ mktemp + kubectl_bin get ns + awk '{print$1}' + local LAST_OUT=/tmp/tmp.xLCXUW7NFC ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.QGNSeTK8yy + local exit_status=0 + local LAST_OUT=/tmp/tmp.OpvpygAB30 ++ seq 0 2 ++ mktemp + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + local LAST_ERR=/tmp/tmp.kiJpx0wVrQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.OpvpygAB30 + cat /tmp/tmp.kiJpx0wVrQ + rm /tmp/tmp.OpvpygAB30 /tmp/tmp.kiJpx0wVrQ + return 0 namespace "cert-manager" deleted namespace "pitr-pxb-18464" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xLCXUW7NFC namespace "pxc-operator" deleted + cat /tmp/tmp.QGNSeTK8yy + rm /tmp/tmp.xLCXUW7NFC /tmp/tmp.QGNSeTK8yy + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.zeEw5oGEcP ++ mktemp + local LAST_ERR=/tmp/tmp.NMQwKY1EX2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zeEw5oGEcP namespace/pxc-operator created + cat /tmp/tmp.NMQwKY1EX2 + rm /tmp/tmp.zeEw5oGEcP /tmp/tmp.NMQwKY1EX2 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.SHKRDIZwc1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bHxAy8tXhb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SHKRDIZwc1 ++ cat /tmp/tmp.bHxAy8tXhb ++ rm /tmp/tmp.SHKRDIZwc1 /tmp/tmp.bHxAy8tXhb ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster2 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.H31T7nVdqU ++ mktemp + local LAST_ERR=/tmp/tmp.or3OmqYP2l + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster2 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.H31T7nVdqU Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster2" modified. + cat /tmp/tmp.or3OmqYP2l + rm /tmp/tmp.H31T7nVdqU /tmp/tmp.or3OmqYP2l + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.PQs8XXlngi ++ mktemp + local LAST_ERR=/tmp/tmp.0ODRvsSXXF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PQs8XXlngi customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.0ODRvsSXXF + rm /tmp/tmp.PQs8XXlngi /tmp/tmp.0ODRvsSXXF + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-rbac.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.qAhPNK5vjE ++ mktemp + local LAST_ERR=/tmp/tmp.DpRTbwKIMb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qAhPNK5vjE clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.DpRTbwKIMb + rm /tmp/tmp.qAhPNK5vjE /tmp/tmp.DpRTbwKIMb + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = "XtrabackupSidecar=true"' - + kubectl_bin apply -f - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e^' ++ mktemp + local LAST_OUT=/tmp/tmp.GgMVt6FQVD ++ mktemp + local LAST_ERR=/tmp/tmp.njggpFIhwW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.GgMVt6FQVD deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.njggpFIhwW + rm /tmp/tmp.GgMVt6FQVD /tmp/tmp.njggpFIhwW + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.PPMGkbPlbA ++ mktemp + local LAST_ERR=/tmp/tmp.s6CqygbJZc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PPMGkbPlbA pod/percona-xtradb-cluster-operator-6fc4ff78b7-6c2dn condition met + cat /tmp/tmp.s6CqygbJZc + rm /tmp/tmp.PPMGkbPlbA /tmp/tmp.s6CqygbJZc + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ head -1 +++ mktemp ++ local LAST_OUT=/tmp/tmp.UQrAijSBqx +++ mktemp ++ local LAST_ERR=/tmp/tmp.OuSWL5Pe6q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UQrAijSBqx ++ cat /tmp/tmp.OuSWL5Pe6q ++ rm /tmp/tmp.UQrAijSBqx /tmp/tmp.OuSWL5Pe6q ++ return 0 + wait_pod percona-xtradb-cluster-operator-6fc4ff78b7-6c2dn 480 pxc-operator + local pod=percona-xtradb-cluster-operator-6fc4ff78b7-6c2dn + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-6fc4ff78b7-6c2dn ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-6fc4ff78b7-6c2dn condition met waiting for pod/percona-xtradb-cluster-operator-6fc4ff78b7-6c2dn to become Ready.Ok + sleep 3 + create_namespace pitr-pxb-531 + local namespace=pitr-pxb-531 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ tail -n1 ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl api-resources ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + desc 'cleaned up old namespaces pitr-pxb-531' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pitr-pxb-531 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pitr-pxb-531 ++ mktemp + xargs kubectl delete ns + awk '{print$1}' + local LAST_OUT=/tmp/tmp.0pOMRzfB6K ++ mktemp + local LAST_ERR=/tmp/tmp.Hcs93lc7fk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-pxb-531 + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.C07KPCfxFa ++ mktemp + local LAST_ERR=/tmp/tmp.1T8TUxlSL2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-pxb-531 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.C07KPCfxFa + cat /tmp/tmp.1T8TUxlSL2 + rm /tmp/tmp.C07KPCfxFa /tmp/tmp.1T8TUxlSL2 + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-pxb-531 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.0pOMRzfB6K + cat /tmp/tmp.Hcs93lc7fk Error from server (NotFound): namespaces "pitr-pxb-531" not found + rm /tmp/tmp.0pOMRzfB6K /tmp/tmp.Hcs93lc7fk + return 1 + : + wait_for_delete namespace/pitr-pxb-531 + local res=namespace/pitr-pxb-531 + echo -n 'waiting for namespace/pitr-pxb-531 to be deleted' waiting for namespace/pitr-pxb-531 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pitr-pxb-531" not found + desc 'create namespace pitr-pxb-531' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pitr-pxb-531 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pitr-pxb-531 ++ mktemp + local LAST_OUT=/tmp/tmp.enX6TLvz3m ++ mktemp + local LAST_ERR=/tmp/tmp.BYYECa8NfX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pitr-pxb-531 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.enX6TLvz3m namespace/pitr-pxb-531 created + cat /tmp/tmp.BYYECa8NfX + rm /tmp/tmp.enX6TLvz3m /tmp/tmp.BYYECa8NfX + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.pVGsCONh5B +++ mktemp ++ local LAST_ERR=/tmp/tmp.l811Qgk5TP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pVGsCONh5B ++ cat /tmp/tmp.l811Qgk5TP ++ rm /tmp/tmp.pVGsCONh5B /tmp/tmp.l811Qgk5TP ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster2 --namespace=pitr-pxb-531 ++ mktemp + local LAST_OUT=/tmp/tmp.3WenssnhAm ++ mktemp + local LAST_ERR=/tmp/tmp.icYahfMXYl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster2 --namespace=pitr-pxb-531 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3WenssnhAm Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster2" modified. + cat /tmp/tmp.icYahfMXYl + rm /tmp/tmp.3WenssnhAm /tmp/tmp.icYahfMXYl + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.KcLyj13ieE ++ mktemp + local LAST_ERR=/tmp/tmp.hVK35a9U7V + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KcLyj13ieE secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.hVK35a9U7V + rm /tmp/tmp.KcLyj13ieE /tmp/tmp.hVK35a9U7V + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.LTebmUlKOb ++ mktemp + local LAST_ERR=/tmp/tmp.upO6NB3tpk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LTebmUlKOb namespace/cert-manager created + cat /tmp/tmp.upO6NB3tpk + rm /tmp/tmp.LTebmUlKOb /tmp/tmp.upO6NB3tpk + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.XWbMylelQd ++ mktemp + local LAST_ERR=/tmp/tmp.AK7YMpPlAT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.XWbMylelQd namespace/cert-manager labeled + cat /tmp/tmp.AK7YMpPlAT + rm /tmp/tmp.XWbMylelQd /tmp/tmp.AK7YMpPlAT + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.VYVq0iuTvU ++ mktemp + local LAST_ERR=/tmp/tmp.uvWeBuPClt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VYVq0iuTvU namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.uvWeBuPClt Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.VYVq0iuTvU /tmp/tmp.uvWeBuPClt + return 0 + '[' '' == 4.10 ']' + sleep 70 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/issuer.yml ++ mktemp + local LAST_OUT=/tmp/tmp.86jD3Q8Ppd ++ mktemp + local LAST_ERR=/tmp/tmp.TYVc6HR1jp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/issuer.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.86jD3Q8Ppd issuer.cert-manager.io/selfsigning-issuer created + cat /tmp/tmp.TYVc6HR1jp + rm /tmp/tmp.86jD3Q8Ppd /tmp/tmp.TYVc6HR1jp + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/cert.yml ++ mktemp + local LAST_OUT=/tmp/tmp.HFoSz2qb7W ++ mktemp + local LAST_ERR=/tmp/tmp.61TvWSsCz3 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/cert.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HFoSz2qb7W certificate.cert-manager.io/tls-minio created + cat /tmp/tmp.61TvWSsCz3 Warning: spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. + rm /tmp/tmp.HFoSz2qb7W /tmp/tmp.61TvWSsCz3 + return 0 + sleep_with_log 25 + local d=25 + log 'sleeping for 25 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:50:27+0000]' sleeping for 25 seconds [2026-04-29T16:50:27+0000] sleeping for 25 seconds + sleep 25 + start_minio tls-minio + deploy_helm pitr-pxb-531 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=tls-minio + local endpoint=http://minio-service:9000 + minio_args=(--version $MINIO_VER --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set "users[0].accessKey=some-access-key" --set "users[0].secretKey=some-secret-key" --set "users[0].policy=consoleAdmin" --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G) + local minio_args + [[ -n tls-minio ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Wed Apr 29 16:50:57 2026 NAMESPACE: pitr-pxb-531 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.pitr-pxb-531.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace pitr-pxb-531 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace pitr-pxb-531 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace pitr-pxb-531 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace pitr-pxb-531 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EyMeo9yvLM +++ mktemp ++ local LAST_ERR=/tmp/tmp.sNlujZrm8m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EyMeo9yvLM ++ cat /tmp/tmp.sNlujZrm8m ++ rm /tmp/tmp.EyMeo9yvLM /tmp/tmp.sNlujZrm8m ++ return 0 + MINIO_POD=minio-service-757b74c96c-6h7z9 + wait_pod minio-service-757b74c96c-6h7z9 + local pod=minio-service-757b74c96c-6h7z9 + local max_retry=480 + local ns= ++ echo minio-service-757b74c96c-6h7z9 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/minio-service-757b74c96c-6h7z9 condition met waiting for pod/minio-service-757b74c96c-6h7z9 to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.NSmARsrazX ++ mktemp + local LAST_ERR=/tmp/tmp.TbaJYQLyZK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NSmARsrazX /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from pitr-pxb-531 namespace + cat /tmp/tmp.TbaJYQLyZK All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: unable to upgrade connection: container aws-cli not found in pod aws-cli_pitr-pxb-531 + rm /tmp/tmp.NSmARsrazX /tmp/tmp.TbaJYQLyZK + return 0 + cluster=pitr + spinup_pxc pitr /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local cluster=pitr + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.WML8AMLMrd ++ mktemp + local LAST_ERR=/tmp/tmp.SM56Oo7upN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WML8AMLMrd secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.SM56Oo7upN + rm /tmp/tmp.WML8AMLMrd /tmp/tmp.SM56Oo7upN + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-pxb-531~ + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' ++ mktemp + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.7ETjgBibGi ++ mktemp + local LAST_ERR=/tmp/tmp.n8JtBvuGzh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.7ETjgBibGi deployment.apps/pxc-client created + cat /tmp/tmp.n8JtBvuGzh + rm /tmp/tmp.7ETjgBibGi /tmp/tmp.n8JtBvuGzh + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local pvc_name= + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-pxb-531~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + kubectl_bin apply -f - + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + local LAST_OUT=/tmp/tmp.vJgLtl5MD9 + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + local LAST_ERR=/tmp/tmp.KT4OpzbZKS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vJgLtl5MD9 perconaxtradbcluster.pxc.percona.com/pitr created + cat /tmp/tmp.KT4OpzbZKS + rm /tmp/tmp.vJgLtl5MD9 /tmp/tmp.KT4OpzbZKS + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.VkST5p0ZtW ++++ mktemp +++ local LAST_ERR=/tmp/tmp.nvu8lxgnBU +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.VkST5p0ZtW +++ cat /tmp/tmp.nvu8lxgnBU +++ rm /tmp/tmp.VkST5p0ZtW /tmp/tmp.nvu8lxgnBU +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.9efmzrtOPC ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fKrGE40wM1 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.9efmzrtOPC +++ cat /tmp/tmp.fKrGE40wM1 +++ rm /tmp/tmp.9efmzrtOPC /tmp/tmp.fKrGE40wM1 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-531 ++ mktemp + local LAST_OUT=/tmp/tmp.whWHu5FbTE ++ mktemp + local LAST_ERR=/tmp/tmp.HqaPRycXbT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-531 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-531 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-531 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.whWHu5FbTE + cat /tmp/tmp.HqaPRycXbT error: no matching resources found + rm /tmp/tmp.whWHu5FbTE /tmp/tmp.HqaPRycXbT + return 1 + true + wait_for_running pitr-proxysql 1 + local name=pitr-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-proxysql-0 + local container=proxysql + set +o xtrace pod/pitr-proxysql-0 condition met waiting for pod/pitr-proxysql-0 to become Ready.Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ echo pitr-pxc-0 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ echo pitr-pxc-1 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc pitr -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.FVZTiwNTGk +++ mktemp ++ local LAST_ERR=/tmp/tmp.9pLKXDNZ4y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FVZTiwNTGk ++ cat /tmp/tmp.9pLKXDNZ4y ++ rm /tmp/tmp.FVZTiwNTGk /tmp/tmp.9pLKXDNZ4y ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TEP1aA6CvP +++ mktemp ++ local LAST_ERR=/tmp/tmp.q6AGBwdjkU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TEP1aA6CvP ++ cat /tmp/tmp.q6AGBwdjkU ++ rm /tmp/tmp.TEP1aA6CvP /tmp/tmp.q6AGBwdjkU ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-sznnl + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3EfFkNPCYH +++ mktemp ++ local LAST_ERR=/tmp/tmp.OiiSPutw8n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3EfFkNPCYH ++ cat /tmp/tmp.OiiSPutw8n ++ rm /tmp/tmp.3EfFkNPCYH /tmp/tmp.OiiSPutw8n ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-sznnl ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VtLtjHwGOg +++ mktemp ++ local LAST_ERR=/tmp/tmp.Kba9nzAMP9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VtLtjHwGOg ++ cat /tmp/tmp.Kba9nzAMP9 ++ rm /tmp/tmp.VtLtjHwGOg /tmp/tmp.Kba9nzAMP9 ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-sznnl ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.EjbdapnVIn/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.EjbdapnVIn/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql /tmp/tmp.EjbdapnVIn/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fO9k47H52q +++ mktemp ++ local LAST_ERR=/tmp/tmp.yV50tFj0l8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fO9k47H52q ++ cat /tmp/tmp.yV50tFj0l8 ++ rm /tmp/tmp.fO9k47H52q /tmp/tmp.yV50tFj0l8 ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-sznnl + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.EjbdapnVIn/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.EjbdapnVIn/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql /tmp/tmp.EjbdapnVIn/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MOSLD8Vbk7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.uhnAkutKg2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MOSLD8Vbk7 ++ cat /tmp/tmp.uhnAkutKg2 ++ rm /tmp/tmp.MOSLD8Vbk7 /tmp/tmp.uhnAkutKg2 ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-sznnl ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.EjbdapnVIn/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.EjbdapnVIn/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql /tmp/tmp.EjbdapnVIn/select-1.sql + is_keyring_plugin_in_use pitr + local cluster=pitr + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + kubectl exec pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + grep -E -o 'early-plugin-load=keyring_\w+.so' + return 1 + run_backup pitr on-pitr-minio + local cluster=pitr + local backup=on-pitr-minio + local name= + '[' -z ']' + name=on-pitr-minio + log 'run backup pxc-backup/on-pitr-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:57:50+0000]' run backup pxc-backup/on-pitr-minio [2026-04-29T16:57:50+0000] run backup pxc-backup/on-pitr-minio + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/on-pitr-minio.yml + kubectl_bin apply -f - + yq eval '.metadata.name="on-pitr-minio"' ++ mktemp + local LAST_OUT=/tmp/tmp.OvecsLkW0G ++ mktemp + local LAST_ERR=/tmp/tmp.BF1cQ0wt8S + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.OvecsLkW0G perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio created + cat /tmp/tmp.BF1cQ0wt8S + rm /tmp/tmp.OvecsLkW0G /tmp/tmp.BF1cQ0wt8S + return 0 + wait_backup on-pitr-minio + local backup=on-pitr-minio + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio to reach Succeeded state...........Succeeded + write_test_data pitr + local cluster=pitr + local config= + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hxzNjBtwQG ++++ mktemp +++ local LAST_ERR=/tmp/tmp.2ftk1U9X5s +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.hxzNjBtwQG +++ cat /tmp/tmp.2ftk1U9X5s +++ rm /tmp/tmp.hxzNjBtwQG /tmp/tmp.2ftk1U9X5s +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MGFu9RNVf2 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1g1NwgcHt0 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.MGFu9RNVf2 +++ cat /tmp/tmp.1g1NwgcHt0 +++ rm /tmp/tmp.MGFu9RNVf2 /tmp/tmp.1g1NwgcHt0 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + desc 'write test data' + set +o xtrace ----------------------------------------------------------------------------------- write test data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ScndIUzZoc +++ mktemp ++ local LAST_ERR=/tmp/tmp.yao8Q6nf6n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ScndIUzZoc ++ cat /tmp/tmp.yao8Q6nf6n ++ rm /tmp/tmp.ScndIUzZoc /tmp/tmp.yao8Q6nf6n ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-sznnl ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' '-h pitr-proxysql -uroot -proot_password' + local 'command=INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Og2wFWLM5Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.zLX3uYRlSI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Og2wFWLM5Y ++ cat /tmp/tmp.zLX3uYRlSI ++ rm /tmp/tmp.Og2wFWLM5Y /tmp/tmp.zLX3uYRlSI ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-sznnl ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep_with_log 30 + local d=30 + log 'sleeping for 30 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:58:28+0000]' sleeping for 30 seconds [2026-04-29T16:58:28+0000] sleeping for 30 seconds + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DGvS0QJXYT +++ mktemp ++ local LAST_ERR=/tmp/tmp.7dp0vJOi5t ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DGvS0QJXYT ++ cat /tmp/tmp.7dp0vJOi5t ++ rm /tmp/tmp.DGvS0QJXYT /tmp/tmp.7dp0vJOi5t ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-sznnl ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.EjbdapnVIn/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.EjbdapnVIn/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql /tmp/tmp.EjbdapnVIn/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OiDkItvgtr +++ mktemp ++ local LAST_ERR=/tmp/tmp.6BxWzKn4uR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OiDkItvgtr ++ cat /tmp/tmp.6BxWzKn4uR ++ rm /tmp/tmp.OiDkItvgtr /tmp/tmp.6BxWzKn4uR ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-sznnl + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.EjbdapnVIn/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.EjbdapnVIn/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql /tmp/tmp.EjbdapnVIn/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jEEU843JsB +++ mktemp ++ local LAST_ERR=/tmp/tmp.oedoGfLips ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jEEU843JsB ++ cat /tmp/tmp.oedoGfLips ++ rm /tmp/tmp.jEEU843JsB /tmp/tmp.oedoGfLips ++ return 0 + client_pod=pxc-client-67fc4995bb-sznnl + wait_pod pxc-client-67fc4995bb-sznnl + local pod=pxc-client-67fc4995bb-sznnl + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-sznnl ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.EjbdapnVIn/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.EjbdapnVIn/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql /tmp/tmp.EjbdapnVIn/select-2.sql ++ is_keyring_plugin_in_use pitr ++ local cluster=pitr ++ [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] ++ kubectl exec pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ grep -E -o 'early-plugin-load=keyring_\w+.so' ++ return 1 + '[' '' ']' + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ echo -n pass1 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczE= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczE= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.rsT4CX0ATu ++ mktemp + local LAST_ERR=/tmp/tmp.NvNdRhS5sh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rsT4CX0ATu secret/my-cluster-secrets patched + cat /tmp/tmp.NvNdRhS5sh + rm /tmp/tmp.rsT4CX0ATu /tmp/tmp.NvNdRhS5sh + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mm8o2z36ke +++ mktemp ++ local LAST_ERR=/tmp/tmp.nyGOvQXtvP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mm8o2z36ke ++ cat /tmp/tmp.nyGOvQXtvP ++ rm /tmp/tmp.mm8o2z36ke /tmp/tmp.nyGOvQXtvP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SMgHK9WIqW +++ mktemp ++ local LAST_ERR=/tmp/tmp.kGICXxy4fL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SMgHK9WIqW ++ cat /tmp/tmp.kGICXxy4fL ++ rm /tmp/tmp.SMgHK9WIqW /tmp/tmp.kGICXxy4fL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hnlglzGnGc +++ mktemp ++ local LAST_ERR=/tmp/tmp.RdAt067t9L ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hnlglzGnGc ++ cat /tmp/tmp.RdAt067t9L ++ rm /tmp/tmp.hnlglzGnGc /tmp/tmp.RdAt067t9L ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3U2qG27l7y +++ mktemp ++ local LAST_ERR=/tmp/tmp.K4YTKtSWk6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3U2qG27l7y ++ cat /tmp/tmp.K4YTKtSWk6 ++ rm /tmp/tmp.3U2qG27l7y /tmp/tmp.K4YTKtSWk6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.a3UbeLvBhO +++ mktemp ++ local LAST_ERR=/tmp/tmp.lu4TuWIRbE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.a3UbeLvBhO ++ cat /tmp/tmp.lu4TuWIRbE ++ rm /tmp/tmp.a3UbeLvBhO /tmp/tmp.lu4TuWIRbE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Kt9gndU3zd +++ mktemp ++ local LAST_ERR=/tmp/tmp.kt6tLRQeJF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Kt9gndU3zd ++ cat /tmp/tmp.kt6tLRQeJF ++ rm /tmp/tmp.Kt9gndU3zd /tmp/tmp.kt6tLRQeJF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SYIbfdnYfl +++ mktemp ++ local LAST_ERR=/tmp/tmp.D5yFNRUeQ3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SYIbfdnYfl ++ cat /tmp/tmp.D5yFNRUeQ3 ++ rm /tmp/tmp.SYIbfdnYfl /tmp/tmp.D5yFNRUeQ3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.L78RjBWIOG +++ mktemp ++ local LAST_ERR=/tmp/tmp.gVmgHavyNJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.L78RjBWIOG ++ cat /tmp/tmp.gVmgHavyNJ ++ rm /tmp/tmp.L78RjBWIOG /tmp/tmp.gVmgHavyNJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BAoSVQnvwv +++ mktemp ++ local LAST_ERR=/tmp/tmp.AwDft0JhOy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BAoSVQnvwv ++ cat /tmp/tmp.AwDft0JhOy ++ rm /tmp/tmp.BAoSVQnvwv /tmp/tmp.AwDft0JhOy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N5bMXgM57Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.lqfKa5HXj6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.N5bMXgM57Z ++ cat /tmp/tmp.lqfKa5HXj6 ++ rm /tmp/tmp.N5bMXgM57Z /tmp/tmp.lqfKa5HXj6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KRmOINEagv +++ mktemp ++ local LAST_ERR=/tmp/tmp.mogxNQu9VK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KRmOINEagv ++ cat /tmp/tmp.mogxNQu9VK ++ rm /tmp/tmp.KRmOINEagv /tmp/tmp.mogxNQu9VK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v0qyjgTCwe +++ mktemp ++ local LAST_ERR=/tmp/tmp.pKSlAsm6jq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v0qyjgTCwe ++ cat /tmp/tmp.pKSlAsm6jq ++ rm /tmp/tmp.v0qyjgTCwe /tmp/tmp.pKSlAsm6jq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MSTTil5ofM +++ mktemp ++ local LAST_ERR=/tmp/tmp.CuyZmk4NcN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MSTTil5ofM ++ cat /tmp/tmp.CuyZmk4NcN ++ rm /tmp/tmp.MSTTil5ofM /tmp/tmp.CuyZmk4NcN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o2UXcz48uO +++ mktemp ++ local LAST_ERR=/tmp/tmp.2rHtjOg11I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o2UXcz48uO ++ cat /tmp/tmp.2rHtjOg11I ++ rm /tmp/tmp.o2UXcz48uO /tmp/tmp.2rHtjOg11I ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EG2VckCYYL +++ mktemp ++ local LAST_ERR=/tmp/tmp.YTs4035xfv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EG2VckCYYL ++ cat /tmp/tmp.YTs4035xfv ++ rm /tmp/tmp.EG2VckCYYL /tmp/tmp.YTs4035xfv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q4KVTyw2vM +++ mktemp ++ local LAST_ERR=/tmp/tmp.YngnOscZyi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.q4KVTyw2vM ++ cat /tmp/tmp.YngnOscZyi ++ rm /tmp/tmp.q4KVTyw2vM /tmp/tmp.YngnOscZyi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XLZKeZnxqc +++ mktemp ++ local LAST_ERR=/tmp/tmp.reBImpAEN0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XLZKeZnxqc ++ cat /tmp/tmp.reBImpAEN0 ++ rm /tmp/tmp.XLZKeZnxqc /tmp/tmp.reBImpAEN0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UDAKz8h5iI +++ mktemp ++ local LAST_ERR=/tmp/tmp.TcNrCcBYby ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UDAKz8h5iI ++ cat /tmp/tmp.TcNrCcBYby ++ rm /tmp/tmp.UDAKz8h5iI /tmp/tmp.TcNrCcBYby ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SaO3b6Ehtl +++ mktemp ++ local LAST_ERR=/tmp/tmp.1t7Qh7Phi6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SaO3b6Ehtl ++ cat /tmp/tmp.1t7Qh7Phi6 ++ rm /tmp/tmp.SaO3b6Ehtl /tmp/tmp.1t7Qh7Phi6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0BgeTl3Ci2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.XbQBY87v7W ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0BgeTl3Ci2 ++ cat /tmp/tmp.XbQBY87v7W ++ rm /tmp/tmp.0BgeTl3Ci2 /tmp/tmp.XbQBY87v7W ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DR36AgaLEz +++ mktemp ++ local LAST_ERR=/tmp/tmp.wJ5w6ct8EP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DR36AgaLEz ++ cat /tmp/tmp.wJ5w6ct8EP ++ rm /tmp/tmp.DR36AgaLEz /tmp/tmp.wJ5w6ct8EP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I14Wo78GEs +++ mktemp ++ local LAST_ERR=/tmp/tmp.KIrlTL8u1s ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.I14Wo78GEs ++ cat /tmp/tmp.KIrlTL8u1s ++ rm /tmp/tmp.I14Wo78GEs /tmp/tmp.KIrlTL8u1s ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TAf9RXwjAo +++ mktemp ++ local LAST_ERR=/tmp/tmp.kHlBRiEvQf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TAf9RXwjAo ++ cat /tmp/tmp.kHlBRiEvQf ++ rm /tmp/tmp.TAf9RXwjAo /tmp/tmp.kHlBRiEvQf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zByetdEdSi +++ mktemp ++ local LAST_ERR=/tmp/tmp.nsJMF3Lv1o ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zByetdEdSi ++ cat /tmp/tmp.nsJMF3Lv1o ++ rm /tmp/tmp.zByetdEdSi /tmp/tmp.nsJMF3Lv1o ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wPuSj8AybQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.iF8xjnqmrg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wPuSj8AybQ ++ cat /tmp/tmp.iF8xjnqmrg ++ rm /tmp/tmp.wPuSj8AybQ /tmp/tmp.iF8xjnqmrg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5be4P3brHI +++ mktemp ++ local LAST_ERR=/tmp/tmp.KgMXIL1k7n ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5be4P3brHI ++ cat /tmp/tmp.KgMXIL1k7n ++ rm /tmp/tmp.5be4P3brHI /tmp/tmp.KgMXIL1k7n ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.479N5U5fwn +++ mktemp ++ local LAST_ERR=/tmp/tmp.pTaLB77Nfz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.479N5U5fwn ++ cat /tmp/tmp.pTaLB77Nfz ++ rm /tmp/tmp.479N5U5fwn /tmp/tmp.pTaLB77Nfz ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.cK8OgX2Ogs ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.F0v25F0NQb +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.cK8OgX2Ogs +++++ cat /tmp/tmp.F0v25F0NQb +++++ rm /tmp/tmp.cK8OgX2Ogs /tmp/tmp.F0v25F0NQb +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.79aCnZkIk3 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.ADyhpgfZVB +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.79aCnZkIk3 +++++ cat /tmp/tmp.ADyhpgfZVB +++++ rm /tmp/tmp.79aCnZkIk3 /tmp/tmp.ADyhpgfZVB +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dMiBVndWXg +++ mktemp ++ local LAST_ERR=/tmp/tmp.8I1V5ctmqO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dMiBVndWXg ++ cat /tmp/tmp.8I1V5ctmqO ++ rm /tmp/tmp.dMiBVndWXg /tmp/tmp.8I1V5ctmqO ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass2 + patch_secret my-cluster-secrets xtrabackup cGFzczI= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.dowZRpjRkd ++ mktemp + local LAST_ERR=/tmp/tmp.KJNajTc7et + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dowZRpjRkd secret/my-cluster-secrets patched + cat /tmp/tmp.KJNajTc7et + rm /tmp/tmp.dowZRpjRkd /tmp/tmp.KJNajTc7et + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PK4brU10Ye +++ mktemp ++ local LAST_ERR=/tmp/tmp.L6fPhhTr0Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PK4brU10Ye ++ cat /tmp/tmp.L6fPhhTr0Y ++ rm /tmp/tmp.PK4brU10Ye /tmp/tmp.L6fPhhTr0Y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CuTwGwne35 +++ mktemp ++ local LAST_ERR=/tmp/tmp.IpakkeRvAk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CuTwGwne35 ++ cat /tmp/tmp.IpakkeRvAk ++ rm /tmp/tmp.CuTwGwne35 /tmp/tmp.IpakkeRvAk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ff9nCInfS2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1ZhdsFqWoI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ff9nCInfS2 ++ cat /tmp/tmp.1ZhdsFqWoI ++ rm /tmp/tmp.ff9nCInfS2 /tmp/tmp.1ZhdsFqWoI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BvbhgDzU2y +++ mktemp ++ local LAST_ERR=/tmp/tmp.sFAtCmXJds ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BvbhgDzU2y ++ cat /tmp/tmp.sFAtCmXJds ++ rm /tmp/tmp.BvbhgDzU2y /tmp/tmp.sFAtCmXJds ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VLxVKdomk9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Gt9fiVWgTv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VLxVKdomk9 ++ cat /tmp/tmp.Gt9fiVWgTv ++ rm /tmp/tmp.VLxVKdomk9 /tmp/tmp.Gt9fiVWgTv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aU9I7OQdnM +++ mktemp ++ local LAST_ERR=/tmp/tmp.gdmBVH8L1T ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aU9I7OQdnM ++ cat /tmp/tmp.gdmBVH8L1T ++ rm /tmp/tmp.aU9I7OQdnM /tmp/tmp.gdmBVH8L1T ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VcVAdQQChC +++ mktemp ++ local LAST_ERR=/tmp/tmp.Fw6V9QcZJL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VcVAdQQChC ++ cat /tmp/tmp.Fw6V9QcZJL ++ rm /tmp/tmp.VcVAdQQChC /tmp/tmp.Fw6V9QcZJL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mAEiVrie5N +++ mktemp ++ local LAST_ERR=/tmp/tmp.CZzMLZ88Rw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mAEiVrie5N ++ cat /tmp/tmp.CZzMLZ88Rw ++ rm /tmp/tmp.mAEiVrie5N /tmp/tmp.CZzMLZ88Rw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K2W2lyHclr +++ mktemp ++ local LAST_ERR=/tmp/tmp.uaT3DBkKfZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.K2W2lyHclr ++ cat /tmp/tmp.uaT3DBkKfZ ++ rm /tmp/tmp.K2W2lyHclr /tmp/tmp.uaT3DBkKfZ ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zKa7tJtl8N +++ mktemp ++ local LAST_ERR=/tmp/tmp.HjNlTiJAc4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zKa7tJtl8N ++ cat /tmp/tmp.HjNlTiJAc4 ++ rm /tmp/tmp.zKa7tJtl8N /tmp/tmp.HjNlTiJAc4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uSQxXgaeYf +++ mktemp ++ local LAST_ERR=/tmp/tmp.QjwRMBB0AM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uSQxXgaeYf ++ cat /tmp/tmp.QjwRMBB0AM ++ rm /tmp/tmp.uSQxXgaeYf /tmp/tmp.QjwRMBB0AM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v5pfsl4Z4G +++ mktemp ++ local LAST_ERR=/tmp/tmp.q6N40Qm2lt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v5pfsl4Z4G ++ cat /tmp/tmp.q6N40Qm2lt ++ rm /tmp/tmp.v5pfsl4Z4G /tmp/tmp.q6N40Qm2lt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PxJ8cg8cdO +++ mktemp ++ local LAST_ERR=/tmp/tmp.E5dbgXeGT6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PxJ8cg8cdO ++ cat /tmp/tmp.E5dbgXeGT6 ++ rm /tmp/tmp.PxJ8cg8cdO /tmp/tmp.E5dbgXeGT6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HUx0g1LurD +++ mktemp ++ local LAST_ERR=/tmp/tmp.ehELee6wWf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HUx0g1LurD ++ cat /tmp/tmp.ehELee6wWf ++ rm /tmp/tmp.HUx0g1LurD /tmp/tmp.ehELee6wWf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t79lrpRZN3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tlp6h3tRIi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.t79lrpRZN3 ++ cat /tmp/tmp.tlp6h3tRIi ++ rm /tmp/tmp.t79lrpRZN3 /tmp/tmp.tlp6h3tRIi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v7R1eq6IiD +++ mktemp ++ local LAST_ERR=/tmp/tmp.tZU5XVjNJ7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v7R1eq6IiD ++ cat /tmp/tmp.tZU5XVjNJ7 ++ rm /tmp/tmp.v7R1eq6IiD /tmp/tmp.tZU5XVjNJ7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F9HRqufxhe +++ mktemp ++ local LAST_ERR=/tmp/tmp.UtTo2ZnNqk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.F9HRqufxhe ++ cat /tmp/tmp.UtTo2ZnNqk ++ rm /tmp/tmp.F9HRqufxhe /tmp/tmp.UtTo2ZnNqk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tRjtTatcNK +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZlghFme13b ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tRjtTatcNK ++ cat /tmp/tmp.ZlghFme13b ++ rm /tmp/tmp.tRjtTatcNK /tmp/tmp.ZlghFme13b ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jm1qB9S9zI +++ mktemp ++ local LAST_ERR=/tmp/tmp.Py6hRXAcU6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jm1qB9S9zI ++ cat /tmp/tmp.Py6hRXAcU6 ++ rm /tmp/tmp.jm1qB9S9zI /tmp/tmp.Py6hRXAcU6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6hHEybz7V5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.u9y4qVOxNd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6hHEybz7V5 ++ cat /tmp/tmp.u9y4qVOxNd ++ rm /tmp/tmp.6hHEybz7V5 /tmp/tmp.u9y4qVOxNd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7qatRt9e9G +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ec1vNfdx8O ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7qatRt9e9G ++ cat /tmp/tmp.Ec1vNfdx8O ++ rm /tmp/tmp.7qatRt9e9G /tmp/tmp.Ec1vNfdx8O ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PHAoTbDMgy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vtgwt8bfgk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PHAoTbDMgy ++ cat /tmp/tmp.Vtgwt8bfgk ++ rm /tmp/tmp.PHAoTbDMgy /tmp/tmp.Vtgwt8bfgk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YVjqUlyeLu +++ mktemp ++ local LAST_ERR=/tmp/tmp.isOIBYCCvT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YVjqUlyeLu ++ cat /tmp/tmp.isOIBYCCvT ++ rm /tmp/tmp.YVjqUlyeLu /tmp/tmp.isOIBYCCvT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SHspSEI4k5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.RqVSIa5cmm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SHspSEI4k5 ++ cat /tmp/tmp.RqVSIa5cmm ++ rm /tmp/tmp.SHspSEI4k5 /tmp/tmp.RqVSIa5cmm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1OPpB4K19Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.IEdu5uRqoy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1OPpB4K19Y ++ cat /tmp/tmp.IEdu5uRqoy ++ rm /tmp/tmp.1OPpB4K19Y /tmp/tmp.IEdu5uRqoy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FRQAQuIRDS +++ mktemp ++ local LAST_ERR=/tmp/tmp.iWkkaVhcYP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FRQAQuIRDS ++ cat /tmp/tmp.iWkkaVhcYP ++ rm /tmp/tmp.FRQAQuIRDS /tmp/tmp.iWkkaVhcYP ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gahMPPPpeS +++ mktemp ++ local LAST_ERR=/tmp/tmp.6PPP3ACrHL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gahMPPPpeS ++ cat /tmp/tmp.6PPP3ACrHL ++ rm /tmp/tmp.gahMPPPpeS /tmp/tmp.6PPP3ACrHL ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.PtkxbFMNeF ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.9j3gnZQ0yf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.PtkxbFMNeF +++++ cat /tmp/tmp.9j3gnZQ0yf +++++ rm /tmp/tmp.PtkxbFMNeF /tmp/tmp.9j3gnZQ0yf +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.PyEAoKnL6I ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.4kbzBcQwVz +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.PyEAoKnL6I +++++ cat /tmp/tmp.4kbzBcQwVz +++++ rm /tmp/tmp.PyEAoKnL6I /tmp/tmp.4kbzBcQwVz +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aIZX5YEbaP +++ mktemp ++ local LAST_ERR=/tmp/tmp.CJHJBQ1HMn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aIZX5YEbaP ++ cat /tmp/tmp.CJHJBQ1HMn ++ rm /tmp/tmp.aIZX5YEbaP /tmp/tmp.CJHJBQ1HMn ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass3 + patch_secret my-cluster-secrets xtrabackup cGFzczM= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczM= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.2aE7BUhZJy ++ mktemp + local LAST_ERR=/tmp/tmp.Xv1uhXBORL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2aE7BUhZJy secret/my-cluster-secrets patched + cat /tmp/tmp.Xv1uhXBORL + rm /tmp/tmp.2aE7BUhZJy /tmp/tmp.Xv1uhXBORL + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.59Su9xqhpC +++ mktemp ++ local LAST_ERR=/tmp/tmp.mrpTVx469T ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.59Su9xqhpC ++ cat /tmp/tmp.mrpTVx469T ++ rm /tmp/tmp.59Su9xqhpC /tmp/tmp.mrpTVx469T ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.g8BRMcmfbU +++ mktemp ++ local LAST_ERR=/tmp/tmp.F7Jzoq1ttQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.g8BRMcmfbU ++ cat /tmp/tmp.F7Jzoq1ttQ ++ rm /tmp/tmp.g8BRMcmfbU /tmp/tmp.F7Jzoq1ttQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lhCVIzmXm5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.72xLpZ7lzw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lhCVIzmXm5 ++ cat /tmp/tmp.72xLpZ7lzw ++ rm /tmp/tmp.lhCVIzmXm5 /tmp/tmp.72xLpZ7lzw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5VFZ7TO1w0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zTOpmjtGGT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5VFZ7TO1w0 ++ cat /tmp/tmp.zTOpmjtGGT ++ rm /tmp/tmp.5VFZ7TO1w0 /tmp/tmp.zTOpmjtGGT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vUIOy9wiPH +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dp384Ju6N3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vUIOy9wiPH ++ cat /tmp/tmp.Dp384Ju6N3 ++ rm /tmp/tmp.vUIOy9wiPH /tmp/tmp.Dp384Ju6N3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eBxbZdHWZw +++ mktemp ++ local LAST_ERR=/tmp/tmp.b0MAvWrxpR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eBxbZdHWZw ++ cat /tmp/tmp.b0MAvWrxpR ++ rm /tmp/tmp.eBxbZdHWZw /tmp/tmp.b0MAvWrxpR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HHdYVeEd6U +++ mktemp ++ local LAST_ERR=/tmp/tmp.eTf92c7a0E ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HHdYVeEd6U ++ cat /tmp/tmp.eTf92c7a0E ++ rm /tmp/tmp.HHdYVeEd6U /tmp/tmp.eTf92c7a0E ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CvfcylmOXU +++ mktemp ++ local LAST_ERR=/tmp/tmp.p8cU2EbZyT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CvfcylmOXU ++ cat /tmp/tmp.p8cU2EbZyT ++ rm /tmp/tmp.CvfcylmOXU /tmp/tmp.p8cU2EbZyT ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.L9uSgMPrmB +++ mktemp ++ local LAST_ERR=/tmp/tmp.zTney5jkJA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.L9uSgMPrmB ++ cat /tmp/tmp.zTney5jkJA ++ rm /tmp/tmp.L9uSgMPrmB /tmp/tmp.zTney5jkJA ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2A5iwcDwH3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jbrVdTyq8F ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2A5iwcDwH3 ++ cat /tmp/tmp.jbrVdTyq8F ++ rm /tmp/tmp.2A5iwcDwH3 /tmp/tmp.jbrVdTyq8F ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nLWyMRe5sz +++ mktemp ++ local LAST_ERR=/tmp/tmp.Zav9tYDZmF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nLWyMRe5sz ++ cat /tmp/tmp.Zav9tYDZmF ++ rm /tmp/tmp.nLWyMRe5sz /tmp/tmp.Zav9tYDZmF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GH9ZmSAz5l +++ mktemp ++ local LAST_ERR=/tmp/tmp.4cqvJKNJSP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GH9ZmSAz5l ++ cat /tmp/tmp.4cqvJKNJSP ++ rm /tmp/tmp.GH9ZmSAz5l /tmp/tmp.4cqvJKNJSP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WSIKkTDhBI +++ mktemp ++ local LAST_ERR=/tmp/tmp.JvLCa3thdV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WSIKkTDhBI ++ cat /tmp/tmp.JvLCa3thdV ++ rm /tmp/tmp.WSIKkTDhBI /tmp/tmp.JvLCa3thdV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h9O4EfYXLl +++ mktemp ++ local LAST_ERR=/tmp/tmp.JKqg9aayq5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h9O4EfYXLl ++ cat /tmp/tmp.JKqg9aayq5 ++ rm /tmp/tmp.h9O4EfYXLl /tmp/tmp.JKqg9aayq5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KZMQE3O9qD +++ mktemp ++ local LAST_ERR=/tmp/tmp.xebfnDQSZt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KZMQE3O9qD ++ cat /tmp/tmp.xebfnDQSZt ++ rm /tmp/tmp.KZMQE3O9qD /tmp/tmp.xebfnDQSZt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wu5SjghADW +++ mktemp ++ local LAST_ERR=/tmp/tmp.TbYqZmkma1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wu5SjghADW ++ cat /tmp/tmp.TbYqZmkma1 ++ rm /tmp/tmp.wu5SjghADW /tmp/tmp.TbYqZmkma1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3NqPY3y9Dg +++ mktemp ++ local LAST_ERR=/tmp/tmp.9snfGrCxeb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3NqPY3y9Dg ++ cat /tmp/tmp.9snfGrCxeb ++ rm /tmp/tmp.3NqPY3y9Dg /tmp/tmp.9snfGrCxeb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ynPy0onU8d +++ mktemp ++ local LAST_ERR=/tmp/tmp.7bou6NFSwg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ynPy0onU8d ++ cat /tmp/tmp.7bou6NFSwg ++ rm /tmp/tmp.ynPy0onU8d /tmp/tmp.7bou6NFSwg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pz26I3MwnY +++ mktemp ++ local LAST_ERR=/tmp/tmp.HyWMsaipgb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pz26I3MwnY ++ cat /tmp/tmp.HyWMsaipgb ++ rm /tmp/tmp.pz26I3MwnY /tmp/tmp.HyWMsaipgb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PLz7YLwCkx +++ mktemp ++ local LAST_ERR=/tmp/tmp.0OdEa1b8yu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PLz7YLwCkx ++ cat /tmp/tmp.0OdEa1b8yu ++ rm /tmp/tmp.PLz7YLwCkx /tmp/tmp.0OdEa1b8yu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EyL4pUGfBN +++ mktemp ++ local LAST_ERR=/tmp/tmp.KlPdvgeD6Q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EyL4pUGfBN ++ cat /tmp/tmp.KlPdvgeD6Q ++ rm /tmp/tmp.EyL4pUGfBN /tmp/tmp.KlPdvgeD6Q ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZaADNJ8P6E +++ mktemp ++ local LAST_ERR=/tmp/tmp.I4DPg656PN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZaADNJ8P6E ++ cat /tmp/tmp.I4DPg656PN ++ rm /tmp/tmp.ZaADNJ8P6E /tmp/tmp.I4DPg656PN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.btmrAWSMJv +++ mktemp ++ local LAST_ERR=/tmp/tmp.SHTG0XPSwH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.btmrAWSMJv ++ cat /tmp/tmp.SHTG0XPSwH ++ rm /tmp/tmp.btmrAWSMJv /tmp/tmp.SHTG0XPSwH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mOmgKB00qk +++ mktemp ++ local LAST_ERR=/tmp/tmp.mmByz17PyW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mOmgKB00qk ++ cat /tmp/tmp.mmByz17PyW ++ rm /tmp/tmp.mOmgKB00qk /tmp/tmp.mmByz17PyW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7JZpzSn6rh +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jhh4Q6HNuF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7JZpzSn6rh ++ cat /tmp/tmp.Jhh4Q6HNuF ++ rm /tmp/tmp.7JZpzSn6rh /tmp/tmp.Jhh4Q6HNuF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.24MIDDj12T +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dske0v4pJe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.24MIDDj12T ++ cat /tmp/tmp.Dske0v4pJe ++ rm /tmp/tmp.24MIDDj12T /tmp/tmp.Dske0v4pJe ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YQkyEuZXdz +++ mktemp ++ local LAST_ERR=/tmp/tmp.xcSWabxD4y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YQkyEuZXdz ++ cat /tmp/tmp.xcSWabxD4y ++ rm /tmp/tmp.YQkyEuZXdz /tmp/tmp.xcSWabxD4y ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.SlEUyo37cP ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.jrBWIPr3nG +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.SlEUyo37cP +++++ cat /tmp/tmp.jrBWIPr3nG +++++ rm /tmp/tmp.SlEUyo37cP /tmp/tmp.jrBWIPr3nG +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.WmdQLWlNbg ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.4dNQIvRtDj +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.WmdQLWlNbg +++++ cat /tmp/tmp.4dNQIvRtDj +++++ rm /tmp/tmp.WmdQLWlNbg /tmp/tmp.4dNQIvRtDj +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VFVd24KzfQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.7wGm7R20jJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VFVd24KzfQ ++ cat /tmp/tmp.7wGm7R20jJ ++ rm /tmp/tmp.VFVd24KzfQ /tmp/tmp.7wGm7R20jJ ++ return 0 + [[ 2 == \2 ]] + echo + compare_kubectl secret/pitr-mysql-init + local resource=secret/pitr-mysql-init + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init.yml + local new_result=/tmp/tmp.EjbdapnVIn/secret_pitr-mysql-init.yml + desc 'compare secret/pitr-mysql-init-' + set +o xtrace ----------------------------------------------------------------------------------- compare secret/pitr-mysql-init- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.32 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k129.yml ']' + version_gt 1.27 ++ echo '1.32 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k127.yml ']' + version_gt 1.24 ++ echo '1.32 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k124.yml ']' + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.32 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.32 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("pitr-pxb-531", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml secret/pitr-mysql-init ++ mktemp + local LAST_OUT=/tmp/tmp.zN3GkeHLXg ++ mktemp + local LAST_ERR=/tmp/tmp.B3YludRSU7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml secret/pitr-mysql-init + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zN3GkeHLXg + cat /tmp/tmp.B3YludRSU7 + rm /tmp/tmp.zN3GkeHLXg /tmp/tmp.B3YludRSU7 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init.yml /tmp/tmp.EjbdapnVIn/secret_pitr-mysql-init.yml + log 'compare_kubectl: secret/pitr-mysql-init OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:08:56+0000]' compare_kubectl: secret/pitr-mysql-init OK [2026-04-29T17:08:56+0000] compare_kubectl: secret/pitr-mysql-init OK + mysql_host=pitr-pxc-0.pitr-pxc.pitr-pxb-531 ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-pxb-531 ++ local host=pitr-pxc-0.pitr-pxc.pitr-pxb-531 +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-pxb-531 -uroot -proot_password' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-pxb-531 -uroot -proot_password' +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.ZIObkAUHUR +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.bx4AWUsp1k ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.ZIObkAUHUR ++++ cat /tmp/tmp.bx4AWUsp1k ++++ rm /tmp/tmp.ZIObkAUHUR /tmp/tmp.bx4AWUsp1k ++++ return 0 +++ client_pod=pxc-client-67fc4995bb-sznnl +++ wait_pod pxc-client-67fc4995bb-sznnl +++ local pod=pxc-client-67fc4995bb-sznnl +++ local max_retry=480 +++ local ns= ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ grep -E '^(pxc|proxysql)$' ++++ echo pxc-client-67fc4995bb-sznnl +++ local container= +++ set +o xtrace pod/pxc-client-67fc4995bb-sznnl condition met waiting for pod/pxc-client-67fc4995bb-sznnl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local gtid=d73e287f-43eb-11f1-83d4-aa4594f923a0:29 ++ [[ ! d73e287f-43eb-11f1-83d4-aa4594f923a0:29 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo d73e287f-43eb-11f1-83d4-aa4594f923a0:29 + gtid=d73e287f-43eb-11f1-83d4-aa4594f923a0:29 + log '[CASE 1]: gtid_executed: d73e287f-43eb-11f1-83d4-aa4594f923a0:29' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:09:07+0000]' '[CASE' '1]:' gtid_executed: d73e287f-43eb-11f1-83d4-aa4594f923a0:29 [2026-04-29T17:09:07+0000] [CASE 1]: gtid_executed: d73e287f-43eb-11f1-83d4-aa4594f923a0:29 + write_data_for_pitr pitr 100503 3 + local cluster=pitr + local start=100503 + local rows=3 ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.8vaPTwSkbn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.koJ5OEry37 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.8vaPTwSkbn +++ cat /tmp/tmp.koJ5OEry37 +++ rm /tmp/tmp.8vaPTwSkbn /tmp/tmp.koJ5OEry37 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.t0QXMG7pCK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.6uSM4e0hN1 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.t0QXMG7pCK +++ cat /tmp/tmp.6uSM4e0hN1 +++ rm /tmp/tmp.t0QXMG7pCK /tmp/tmp.6uSM4e0hN1 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + local sql= + (( i=0 )) + (( i//g' + /usr/bin/sed -e 's///g' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/restore-on-pitr-minio-gtid.yaml + /usr/bin/sed -e 's//d73e287f-43eb-11f1-83d4-aa4594f923a0:29/g' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-pxb-531~ ++ mktemp + local LAST_OUT=/tmp/tmp.sdUSpVqYB5 ++ mktemp + local LAST_ERR=/tmp/tmp.jZHRuOW5cU + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sdUSpVqYB5 perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-gtid created + cat /tmp/tmp.jZHRuOW5cU + rm /tmp/tmp.sdUSpVqYB5 /tmp/tmp.jZHRuOW5cU + return 0 + wait_backup_restore restore-on-pitr-minio-gtid 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Stopping Cluster state 2026-04-29T17:11:49 pxc-restore/restore-on-pitr-minio-gtid state: Stopping Cluster + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace .Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-1" not found + wait_for_delete pod/pitr-pxc-0 + local res=pod/pitr-pxc-0 + echo -n 'waiting for pod/pitr-pxc-0 to be deleted' waiting for pod/pitr-pxc-0 to be deleted+ set +o xtrace ...Error from server (NotFound): pods "pitr-pxc-0" not found + wait_backup_restore restore-on-pitr-minio-gtid 'Point-in-time recovering' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Point-in-time recovering' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Point-in-time recovering state 2026-04-29T17:12:29 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:32 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:35 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:37 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:40 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:43 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:46 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:49 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:52 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:56 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:59 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:02 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:06 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:10 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:13 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:17 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:21 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:24 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:28 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:31 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:35 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:38 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:41 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:44 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:47 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:51 pxc-restore/restore-on-pitr-minio-gtid state: Point-in-time recovering + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ........................................................................................................................2026-04-29T16:48:12.754Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.32.13-gke.1362000"} 2026-04-29T16:48:12.754Z INFO setup Manager starting up {"gitCommit": "9bec550e38b53ea8a42441eabf05aff8b718c288", "gitBranch": "PR-2433-9bec550e", "buildTime": "2026-04-29T14:21:52Z", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} 2026-04-29T16:48:12.754Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "XtrabackupSidecar=true", "enabled": "XtrabackupSidecar=true"} 2026-04-29T16:48:12.757Z INFO setup Registering Components. 2026-04-29T16:48:13.045Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-04-29T16:48:13.046Z INFO setup Starting the Cmd. 2026-04-29T16:48:13.046Z INFO controller-runtime.metrics Starting metrics server 2026-04-29T16:48:13.046Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-04-29T16:48:13.046Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-04-29T16:48:13.046Z INFO controller-runtime.webhook Starting webhook server 2026-04-29T16:48:13.046Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-04-29T16:48:13.046Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-04-29T16:48:13.046Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-04-29T16:48:13.146Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:48:13.182Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:48:13.182Z DEBUG events percona-xtradb-cluster-operator-6fc4ff78b7-6c2dn_912bb59e-fb2f-4448-af57-f31d95eafa09 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"ca9616e2-b89b-4736-97b7-b432ede0ca0e","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1777481293175055009"}, "reason": "LeaderElection"} 2026-04-29T16:48:13.182Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-04-29T16:48:13.182Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-04-29T16:48:13.182Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-04-29T16:48:13.183Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-04-29T16:48:13.283Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-04-29T16:48:13.283Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-04-29T16:48:13.283Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-04-29T16:48:13.283Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-04-29T16:48:13.384Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-04-29T16:48:13.384Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-04-29T16:52:15.836Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "version": "1.20.0"} 2026-04-29T16:52:16.138Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e"} 2026-04-29T16:52:19.182Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e"} 2026-04-29T16:52:19.214Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e"} 2026-04-29T16:52:22.314Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "auto-pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:52:22.331Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:52:22.448Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:52:22.526Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:52:22.579Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:52:22.619Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "pitr-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:52:22.660Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:52:22.770Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bc458e28-e116-4fb1-902d-3fcfc0e2b93e", "object": "pitr-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:52:23.831Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "eb7ad7e2-0787-4251-9734-16609b84d90e", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:52:23.857Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "eb7ad7e2-0787-4251-9734-16609b84d90e", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:53:40.030Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba", "user": "operator"} 2026-04-29T16:53:40.093Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba", "user": "monitor"} 2026-04-29T16:53:40.163Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba"} 2026-04-29T16:53:40.210Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba"} 2026-04-29T16:53:40.250Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba", "user": "xtrabackup"} 2026-04-29T16:53:40.305Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba"} 2026-04-29T16:53:40.344Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba", "user": "replication"} 2026-04-29T16:53:40.352Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9895574-4339-4a83-a182-f2270aba54ba", "err": "get primary pxc pod: not found"} 2026-04-29T16:53:45.457Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "791cea58-6bf1-4de3-9aa8-7196f6ff7c10", "err": "get primary pxc pod: not found"} 2026-04-29T16:53:50.563Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "cf0662e9-5bbf-46f6-aea0-b0737ed6b99b", "err": "get primary pxc pod: not found"} 2026-04-29T16:53:55.675Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d22ae79e-a986-42dc-a7c0-527aea93a42f", "err": "get primary pxc pod: not found"} 2026-04-29T16:56:21.619Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6330b780-16f3-4210-9beb-db5dfff96f52", "user": "root"} 2026-04-29T16:56:21.684Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6330b780-16f3-4210-9beb-db5dfff96f52", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}"} 2026-04-29T16:56:21.743Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6330b780-16f3-4210-9beb-db5dfff96f52", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T16:56:21.793Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6330b780-16f3-4210-9beb-db5dfff96f52", "new version": "8.0.43-34.1"} 2026-04-29T16:56:22.434Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "c65fa5e5-7ee6-47af-bf6c-912518709b61", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}", "hashChanged": false, "metaChanged": true}   &v1.Service{    TypeMeta: v1.TypeMeta{ -  Kind: "Service", +  Kind: "", -  APIVersion: "v1", +  APIVersion: "",    },    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "d7367027-3384-4e29-85a9-8a4dafe1ca7d", +  UID: "", -  ResourceVersion: "1777481781738319013", +  ResourceVersion: "",    Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:56:21 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil,    Annotations: map[string]string{ -  "cloud.google.com/neg": `{"ingress":true}`,    "percona.com/last-config-hash": "eyJwb3J0cyI6W3sibmFtZSI6Imh0dHAiLCJwb3J0Ijo4MDgwLCJ0YXJnZXRQb3J0"...,    },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "v1", -  Time: s"2026-04-29 16:56:21 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.ServiceSpec{    Ports: []v1.ServicePort{    {    Name: "http", -  Protocol: "TCP", +  Protocol: "",    AppProtocol: nil,    Port: 8080,    TargetPort: intstr.IntOrString{    Type: 0, -  IntVal: 8080, +  IntVal: 0,    StrVal: "",    },    NodePort: 0,    },    },    Selector: {"app.kubernetes.io/component": "pitr", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    ClusterIP: "34.118.235.119", -  ClusterIPs: []string{"34.118.235.119"}, +  ClusterIPs: nil,    Type: "ClusterIP",    ExternalIPs: nil, -  SessionAffinity: "None", +  SessionAffinity: "",    LoadBalancerIP: "",    LoadBalancerSourceRanges: nil,    ... // 3 identical fields    PublishNotReadyAddresses: false,    SessionAffinityConfig: nil, -  IPFamilies: []v1.IPFamily{"IPv4"}, +  IPFamilies: nil, -  IPFamilyPolicy: &"SingleStack", +  IPFamilyPolicy: nil,    AllocateLoadBalancerNodePorts: nil,    LoadBalancerClass: nil, -  InternalTrafficPolicy: &"Cluster", +  InternalTrafficPolicy: nil,    TrafficDistribution: nil,    },    Status: {},   } 2026-04-29T16:56:23.340Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6330b780-16f3-4210-9beb-db5dfff96f52"} 2026-04-29T16:56:29.656Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6a4adad0-8640-4b26-b52f-1734e9c3fa48"} 2026-04-29T16:56:35.133Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "575679e1-3926-4d58-9332-b72b59a52180"} 2026-04-29T16:56:40.552Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "5e902640-8461-42bf-92e3-473d7234f534"} 2026-04-29T16:56:45.745Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a1cd7785-60d6-4436-8db2-34f2d45094e3"} 2026-04-29T16:56:50.750Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "3bcc8f19-ce6a-47a0-807e-fc35b3ac7969"} 2026-04-29T16:56:56.236Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ab6b761c-a501-4062-ba99-f19b18a3f57d"} 2026-04-29T16:57:01.535Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a24eaa72-fe54-43bb-98e8-049b853c9778"} 2026-04-29T16:57:06.846Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "8a6e1f85-c96f-42f6-b2bd-8d61c724d64f"} 2026-04-29T16:57:12.445Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "580c44b8-97dd-4b48-a8cd-0b324f659c0a"} 2026-04-29T16:57:17.550Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "104e2a49-3d26-4435-ae86-efe40c98ed20"} 2026-04-29T16:57:22.537Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "83046869-cfce-4bf6-84db-6cf9e24d31ae"} 2026-04-29T16:57:28.335Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "3cc7b8fc-3691-4e32-a514-fe15c30fcf6e"} 2026-04-29T16:57:33.349Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ab055894-2bb9-447d-8228-d5e70ca0f885"} 2026-04-29T16:57:38.660Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "10d88791-534e-48d4-9432-8f4686922894"} 2026-04-29T16:57:44.138Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b9b28bc0-7fe4-437c-a6f7-36994d1db279"} 2026-04-29T16:57:49.447Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "eafe2bd3-f6b7-4e48-965a-21275f297a66"} 2026-04-29T16:57:52.846Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "d9ede130-ebaa-4358-b4b9-d688af6b9f2e", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:57:52.904Z INFO Created a new backup job {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "d9ede130-ebaa-4358-b4b9-d688af6b9f2e", "namespace": "pitr-pxb-531", "name": "xb-on-pitr-minio"} 2026-04-29T16:57:54.561Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "34b28890-6b3d-4060-9b5b-f6768f35f29a"} 2026-04-29T16:57:57.905Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "49799028-ea89-4c45-ac4a-63f32771cda9", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:57:57.977Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "1095f12c-5d48-48f4-9333-c69c081f19ec", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:00.139Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "f29277dc-ca92-49d4-8ad3-50cc468e4a9b"} 2026-04-29T16:58:03.021Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "886d46a1-07b9-4ff9-95fc-b75cf92578ef", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:05.432Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "953d6ab5-7b5c-4602-b0ba-afc04c49f191"} 2026-04-29T16:58:08.064Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "896bb12c-3bfd-4664-986a-83fbeb6c5414", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:08.110Z INFO Backup succeeded {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "896bb12c-3bfd-4664-986a-83fbeb6c5414", "job": "xb-on-pitr-minio"} 2026-04-29T16:58:08.110Z DEBUG Removing binlog gap file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "896bb12c-3bfd-4664-986a-83fbeb6c5414", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-d665d6869-gcl2n"} 2026-04-29T16:58:09.650Z DEBUG Removing binlog timeline file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "896bb12c-3bfd-4664-986a-83fbeb6c5414", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-d665d6869-gcl2n"} 2026-04-29T16:58:09.925Z DEBUG Removing mysql-init secret {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "on-pitr-minio", "reconcileID": "896bb12c-3bfd-4664-986a-83fbeb6c5414", "job": "xb-on-pitr-minio", "secret": "pitr-mysql-init"} 2026-04-29T16:58:10.532Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "4bf8ab32-16ac-4dd8-8829-72a3dae75877"} 2026-04-29T16:58:15.944Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "478057e5-45a7-43f7-84ee-682137383d45"} 2026-04-29T16:58:21.754Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "0b82b335-fa68-4563-baab-6146ad08780a"} 2026-04-29T16:58:26.728Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "60935f78-4eb4-4887-b72f-9c71058ac819", "latest": "2026-04-29 16:58:23 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:58:27.645Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "60935f78-4eb4-4887-b72f-9c71058ac819"} 2026-04-29T16:58:33.603Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26ea4a83-e35c-44fe-8061-23558505efc8"} 2026-04-29T16:58:39.559Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "48be575e-5875-48b5-bcc6-7bac9764b8af"} 2026-04-29T16:58:45.140Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d56da1cb-248e-4b81-a393-4caebefd0602"} 2026-04-29T16:58:51.336Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "02fe6da4-37d3-4b97-987d-609703f7d7aa"} 2026-04-29T16:58:56.940Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b55c4fd2-2d43-48c3-9ee1-7db12bbd9f1d"} 2026-04-29T16:59:02.643Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "80c279f8-9271-4425-ab15-875a3b6bf2a2"} 2026-04-29T16:59:08.057Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ed53ccf0-371e-4a55-a02b-620d6674fc1c"} 2026-04-29T16:59:13.931Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "045eb60e-129f-40bf-bbd5-656cd8cecfd6"} 2026-04-29T16:59:19.841Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "3b6a2dd5-fe46-47a9-8b99-fa6fe06a9c8b"} 2026-04-29T16:59:24.054Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "43ebcc69-6e84-4e56-bbf6-8e58a5597f86", "latest": "2026-04-29 16:59:19 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T16:59:25.129Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "43ebcc69-6e84-4e56-bbf6-8e58a5597f86"} 2026-04-29T16:59:30.084Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd5397e8-bb48-40d6-9245-e0e08a0d1df2", "user": "xtrabackup"} 2026-04-29T16:59:30.107Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd5397e8-bb48-40d6-9245-e0e08a0d1df2", "user": "xtrabackup"} 2026-04-29T16:59:30.128Z INFO MySQL init secret created {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd5397e8-bb48-40d6-9245-e0e08a0d1df2", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T16:59:30.147Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd5397e8-bb48-40d6-9245-e0e08a0d1df2", "user": "xtrabackup"} 2026-04-29T16:59:30.168Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd5397e8-bb48-40d6-9245-e0e08a0d1df2", "user": "xtrabackup"} 2026-04-29T16:59:30.174Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd5397e8-bb48-40d6-9245-e0e08a0d1df2", "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258"} 2026-04-29T16:59:30.177Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd5397e8-bb48-40d6-9245-e0e08a0d1df2", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "19913d33-4030-4351-9d8e-6a7dd7c7a98b", +  UID: "", -  ResourceVersion: "1777481779841679024", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiIwNzVkOTU0MzY1MGUwMjQ5Y2E4NzE5Yjc2OTYyYTM1NCIsInBlcmNvbmEuY29tL3NzbC1pbnRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:52:22 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:56:19 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc", +  "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258",    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354",    "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: nil,    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "grpc",    HostPort: 0,    ContainerPort: 6450, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: nil,    Env: []v1.EnvVar{    {    Name: "POD_NAMESPACE",    Value: "",    ValueFrom: &v1.EnvVarSource{    FieldRef: &v1.ObjectFieldSelector{ -  APIVersion: "v1", +  APIVersion: "",    FieldPath: "metadata.namespace",    },    ResourceFieldRef: nil,    ConfigMapKeyRef: nil,    ... // 2 identical fields    },    },    {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}},    {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-7684d858bc", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-7684d858bc", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T16:59:31.293Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a48938b0-ef9f-4dba-a547-028bb269676a"} 2026-04-29T17:00:34.581Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b22211ba-bf71-4a89-86bb-42e4e0495613", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:00:40.552Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "5e83a965-e920-48a8-a01f-e66300cd2f92", "latest": "2026-04-29 17:00:40 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:01:32.052Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d8d04fdc-7e28-49f7-ae64-292b6df35178", "err": "failed to connect to pod pitr-pxc-0: dial tcp 10.15.2.60:33062: connect: connection refused"} 2026-04-29T17:01:37.759Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a59a1209-b266-48a1-b74f-fd1a1abece90", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:01:38.348Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a59a1209-b266-48a1-b74f-fd1a1abece90", "latest": "2026-04-29 17:00:38 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:01:43.454Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a0199502-45a2-447d-96f4-692f0e71e2d3", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:01:49.122Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d2236070-26c6-4a6b-a4a5-e26c8a9eae6a", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:01:54.825Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "f507bde8-6487-4aaf-8d68-63dd6f4081bd", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:02:00.320Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "0155cac7-bff5-424f-b0a3-66ebe672bba0", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:02:06.022Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b2d6b377-6b52-48b0-a7b1-ccf412523a71", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:02:11.700Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "aa4af5ff-e3b5-4ed3-b412-613df3426095", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:02:17.386Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "5c2b6a6b-92e0-4254-b780-6c105e42a3a8", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:02:26.760Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "22136732-8d37-4d7b-9eb3-33dca772b7cd"} 2026-04-29T17:02:32.130Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "52729b25-d3c1-484d-9426-12acdf1af90a"} 2026-04-29T17:02:35.143Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "user": "xtrabackup"} 2026-04-29T17:02:35.169Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "user": "xtrabackup"} 2026-04-29T17:02:35.199Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T17:02:35.222Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "user": "xtrabackup"} 2026-04-29T17:02:35.245Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "user": "xtrabackup"} 2026-04-29T17:02:35.255Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "last-applied-secret": "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d"} 2026-04-29T17:02:35.258Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "19913d33-4030-4351-9d8e-6a7dd7c7a98b", +  UID: "", -  ResourceVersion: "1777482138119279024", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:59:30 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:02:18 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258", +  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354",    "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: nil,    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "grpc",    HostPort: 0,    ContainerPort: 6450, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: nil,    Env: []v1.EnvVar{    {    Name: "POD_NAMESPACE",    Value: "",    ValueFrom: &v1.EnvVarSource{    FieldRef: &v1.ObjectFieldSelector{ -  APIVersion: "v1", +  APIVersion: "",    FieldPath: "metadata.namespace",    },    ResourceFieldRef: nil,    ConfigMapKeyRef: nil,    ... // 2 identical fields    },    },    {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}},    {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-5b749b6f8", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-5b749b6f8", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T17:02:35.973Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11", "latest": "2026-04-29 17:00:40 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:02:37.962Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "26967267-2804-4eeb-b349-9610b49ebf11"} 2026-04-29T17:03:35.291Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "1c1c5e8f-a95d-4466-bb57-346abcfd1816", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:03:37.551Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "e18b65a3-7fd1-44ee-9659-27ac23a51e08", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:03:39.950Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "706fb96b-ed2b-43ba-afba-5d71188f1e88", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:03:42.488Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "88f47353-f87c-48dd-b386-b6e01f31e9ed", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp 10.15.0.75:33062: connect: connection refused"} 2026-04-29T17:03:48.476Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "f5a24e65-96ea-4d77-8f05-c75745deb7a2", "latest": "2026-04-29 17:03:45 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:04:46.380Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d5880714-ccdd-4d30-8b55-0c8a52331767", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:04:51.886Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "004b57c6-25ad-4c3f-ad6d-f118e8560ff8", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:04:57.365Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "f2dc846f-173b-4cda-8d01-5328a8736962", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:05:02.840Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "69a6c8fb-0555-4dc7-81b9-456d22b5a2d6", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:05:08.603Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "170c2998-1fdb-460b-8195-008cb8f8c630", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:05:14.076Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "afc279f3-1b8a-467b-b3a0-55b63791a478", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:05:19.593Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "be7d0656-49b9-4635-b2ab-6f16d45f7e80", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:05:25.333Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "127c9d26-8bf9-464d-b362-ce611be67395", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:05:34.791Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a46263f7-67f2-4586-bcd7-d050b32680d2"} 2026-04-29T17:05:40.143Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "dcb61f43-2f35-4bc8-a3c2-bc8275928032"} 2026-04-29T17:05:45.735Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "cd246cc8-937a-44c4-b6d9-7f53d94d1689"} 2026-04-29T17:05:51.189Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "969e53da-caca-4c2b-9817-ebf5845b098a", "user": "xtrabackup"} 2026-04-29T17:05:51.212Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "969e53da-caca-4c2b-9817-ebf5845b098a", "user": "xtrabackup"} 2026-04-29T17:05:51.236Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "969e53da-caca-4c2b-9817-ebf5845b098a", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T17:05:51.261Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "969e53da-caca-4c2b-9817-ebf5845b098a", "user": "xtrabackup"} 2026-04-29T17:05:51.282Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "969e53da-caca-4c2b-9817-ebf5845b098a", "user": "xtrabackup"} 2026-04-29T17:05:51.288Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "969e53da-caca-4c2b-9817-ebf5845b098a", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27"} 2026-04-29T17:05:51.291Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "969e53da-caca-4c2b-9817-ebf5845b098a", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "19913d33-4030-4351-9d8e-6a7dd7c7a98b", +  UID: "", -  ResourceVersion: "1777482330431535024", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiJkZjhjYmI5YmQ2NzM1MmMxMDIyOGY0ODcwNjkwZmU4MDIwYzE5MzNjNjkxMWY3Yjk0NmM0N2UyYjc2MDgwYjI3IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:02:35 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:30 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: v1.ObjectMeta{    ... // 9 identical fields    DeletionGracePeriodSeconds: nil,    Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...},    Annotations: map[string]string{    "kubectl.kubernetes.io/default-container": "pxc",    "last-applied-secret": strings.Join({ -  "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d", +  "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27",    }, ""),    "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79",    "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354",    "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35",    },    OwnerReferences: nil,    Finalizers: nil,    ManagedFields: nil,    },    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: nil,    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "grpc",    HostPort: 0,    ContainerPort: 6450, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: nil,    Env: []v1.EnvVar{    {    Name: "POD_NAMESPACE",    Value: "",    ValueFrom: &v1.EnvVarSource{    FieldRef: &v1.ObjectFieldSelector{ -  APIVersion: "v1", +  APIVersion: "",    FieldPath: "metadata.namespace",    },    ResourceFieldRef: nil,    ConfigMapKeyRef: nil,    ... // 2 identical fields    },    },    {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}},    {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-5cfcf4f596", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-5cfcf4f596", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T17:05:52.095Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bd788130-e8f2-4222-8e3b-b0bd3418ee2a"} 2026-04-29T17:06:39.948Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "dda35eac-6be0-4a05-b941-96453322694f", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:41.674Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "db340fa7-6973-4cec-85f7-5eb91927704c", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:43.759Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "06643056-d355-46e7-bf17-44fe51a4b891", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:46.193Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6c40b563-68d5-4bec-b135-25a3da1ca24b", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:48.554Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a9530303-0d9a-4572-aad8-f565c0de4e2e", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:50.495Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "9cf6dbab-f1d4-4509-97f5-087968166e0e", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:53.125Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "c2e32b87-7890-4743-9aff-340f10416786", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:55.652Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "f72fe8a9-7406-43ef-b299-c968c6085401", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:06:58.965Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ae80ebe4-5984-4e76-b170-231b839b64d3", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:09.539Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "890b7a77-efa9-4ad4-b4b1-5edd4a4ba98e", "latest": "2026-04-29 17:07:06 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:07:50.170Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "be4349e3-4fbc-414d-9931-515fb3149b2e", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:07:56.598Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "06f813ca-fd99-4be5-ad2d-8e737bcf5de5", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:08:02.128Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "1381cbff-ce9f-413e-ba44-e6304de69f5f", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:07.637Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "1895e0b8-aefa-4d9f-8094-2095740a0f2c", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:13.138Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "1f0b6011-3418-4a2a-aa96-0c7ce874d9e5", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:18.709Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "dcc46352-759f-468f-810e-eb4216f428d1", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:24.205Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "90db06b2-1f32-45dc-9d6e-fc9cfcd621e2", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:29.901Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b5a906e7-9f75-435c-b53b-8d5b7072f05a", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:35.598Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "e0763d39-e8b7-4759-aeef-665aa893edd1", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:41.348Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "92e26201-6aa7-475e-993e-2140d6680768", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-531.svc.cluster.local"} 2026-04-29T17:08:50.458Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "3ffb1de4-c9a2-4f57-b92c-cce5cd0b0459"} 2026-04-29T17:08:55.717Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "0183bd42-b692-4cb8-ab78-e3b2ad2cea72"} 2026-04-29T17:09:01.252Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "37ac6f88-f350-45b1-887b-7c5b56a6504e"} 2026-04-29T17:09:06.834Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "011af1f4-5605-4828-a52f-a72fa8544c89"} 2026-04-29T17:09:12.932Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "c1ddc4e6-8916-4dd3-b9ef-cb86054b48a9"} 2026-04-29T17:09:18.562Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "606b9087-3002-413b-929d-effee8209891"} 2026-04-29T17:09:25.234Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "57319199-2d31-478f-808a-75d2d5ee4b4e"} 2026-04-29T17:09:31.034Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a73e092f-4d51-482f-8543-884b809c95e7"} 2026-04-29T17:09:36.436Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6fa55845-0a86-4757-b35d-cec33ca0ac56"} 2026-04-29T17:09:42.738Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "09bb50d9-4089-4e0f-8366-8fadffd4b6d4"} 2026-04-29T17:09:48.638Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a0beb69c-f616-40b4-bdcd-2839ff87cce8"} 2026-04-29T17:09:54.610Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d9604426-6d26-48b8-8247-d1dcc5a8975f"} 2026-04-29T17:09:59.142Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "0fc9af18-e501-46f9-bdd7-f4a2719017a1", "latest": "2026-04-29 17:09:54 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:10:00.338Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "0fc9af18-e501-46f9-bdd7-f4a2719017a1"} 2026-04-29T17:10:06.146Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "acd6c5ea-c499-4048-9f32-38ef7aa904ae"} 2026-04-29T17:10:12.235Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "23cc6858-6ea5-43af-8023-822407abe2f5"} 2026-04-29T17:10:17.931Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "02a06cd0-45a5-4998-8a4b-f4c18ce3bb1b"} 2026-04-29T17:10:23.581Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "5b262b7e-cc0c-45cf-ba2f-157d9557c70f"} 2026-04-29T17:10:29.131Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ca9f65ea-fa77-4a4f-a4ab-b32839ca921b"} 2026-04-29T17:10:35.053Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a1dcd3a2-2871-404a-a799-05b3856da97d"} 2026-04-29T17:10:40.550Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "69a2c32e-6352-4f06-8ae4-dd843b2bd832"} 2026-04-29T17:10:46.538Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ff6a1e10-aade-4847-91d5-479ca56ce602"} 2026-04-29T17:10:52.420Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "057ce2d2-3564-4aed-bf3f-4f8bcb54544b"} 2026-04-29T17:10:58.059Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "467651ec-eaa6-424e-82b1-8711fda985a9"} 2026-04-29T17:11:03.961Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "bf9411d1-8e2f-48de-b1c2-036a6126b367"} 2026-04-29T17:11:09.845Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "47323d5c-20f6-407b-829a-7d074368b79a"} 2026-04-29T17:11:15.440Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "2897091c-177a-41c5-991d-778a8d90df20"} 2026-04-29T17:11:21.347Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "54eb7483-5dfb-4d75-8e09-20d03d45b5ea"} 2026-04-29T17:11:26.545Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "e4dae693-cabc-4e34-9923-417325b0c21e"} 2026-04-29T17:11:32.735Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "eeb682b2-b07f-436f-9f5d-345c7be9b21d"} 2026-04-29T17:11:38.831Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "c9ab801d-6098-409d-9c2c-7d2ebd4c9c9c"} 2026-04-29T17:11:44.842Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "1f7acfa6-6b86-4f99-acda-d67fa56bec6c"} 2026-04-29T17:11:46.813Z INFO stopping cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7a90d21d-f58a-4c0e-a59a-5e82cde0d903", "cluster": "pitr"} 2026-04-29T17:11:46.938Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "2352777d-996d-4c11-859f-9252683b4265"} 2026-04-29T17:11:46.939Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "2352777d-996d-4c11-859f-9252683b4265", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "19913d33-4030-4351-9d8e-6a7dd7c7a98b", +  UID: "", -  ResourceVersion: "1777482524707407024", +  ResourceVersion: "", -  Generation: 4, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:05:51 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:08:44 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &3, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: nil,    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "grpc",    HostPort: 0,    ContainerPort: 6450, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: nil,    Env: []v1.EnvVar{    {    Name: "POD_NAMESPACE",    Value: "",    ValueFrom: &v1.EnvVarSource{    FieldRef: &v1.ObjectFieldSelector{ -  APIVersion: "v1", +  APIVersion: "",    FieldPath: "metadata.namespace",    },    ResourceFieldRef: nil,    ConfigMapKeyRef: nil,    ... // 2 identical fields    },    },    {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}},    {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 4, +  ObservedGeneration: 0, -  Replicas: 3, +  Replicas: 0, -  ReadyReplicas: 3, +  ReadyReplicas: 0, -  CurrentReplicas: 3, +  CurrentReplicas: 0, -  UpdatedReplicas: 3, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-55945d5dd", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-55945d5dd", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 3, +  AvailableReplicas: 0,    },   } 2026-04-29T17:11:47.001Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "2352777d-996d-4c11-859f-9252683b4265", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "a16d388a-1d1c-4e2f-9fde-9f7a21700583", +  UID: "", -  ResourceVersion: "1777481578252815022", +  ResourceVersion: "", -  Generation: 1, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:52:22 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 16:52:58 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &2, +  Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 1, +  ObservedGeneration: 0, -  Replicas: 2, +  Replicas: 0, -  ReadyReplicas: 2, +  ReadyReplicas: 0, -  CurrentReplicas: 2, +  CurrentReplicas: 0, -  UpdatedReplicas: 2, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-69b7cdbb99", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-69b7cdbb99", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 2, +  AvailableReplicas: 0,    },   } 2026-04-29T17:11:49.418Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "2352777d-996d-4c11-859f-9252683b4265", "error": "exec binlog collector pod pitr-pitr-d665d6869-gcl2n: failed to execute command in pod: pods \"pitr-pitr-d665d6869-gcl2n\" not found", "errorVerbose": "pods \"pitr-pitr-d665d6869-gcl2n\" not found\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.CheckPITRErrors\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:73\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:459\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-d665d6869-gcl2n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.CheckPITRErrors\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:78\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:459\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:12:27.814Z INFO starting restore {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "314cd39a-5b81-41a5-b2ac-7fcc5d4ed15e", "cluster": "pitr", "backup": "on-pitr-minio"} 2026-04-29T17:12:27.926Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9f4270cd-7753-479f-930f-cad8ec514b03", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:32.960Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "6a1c76b5-b348-45de-9b2e-97482f5001f7", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:37.981Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f0b1f939-c186-4821-ae6d-5f12f103f61f", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:42.999Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "89049934-73dd-4784-bba6-5f9af6c95b12", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:48.018Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "54594999-a75b-45a7-b166-8aec5a1a63c2", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:53.036Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9211ec6d-8b9c-4581-819c-e31e4870ee54", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:58.057Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f5fd192e-ec45-4051-9bba-6bdef292363d", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:13:03.078Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7a0b6f96-96ee-4640-aa31-13ff5a50444f", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:03.078Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7a0b6f96-96ee-4640-aa31-13ff5a50444f", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:03.079Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7a0b6f96-96ee-4640-aa31-13ff5a50444f", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:03.253Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "22c19580-f8f8-4b87-a5d7-49676f745cdd", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "19913d33-4030-4351-9d8e-6a7dd7c7a98b", +  UID: "", -  ResourceVersion: "1777482746132159024", +  ResourceVersion: "", -  Generation: 5, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:11:46 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:12:26 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &1,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: nil,    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "grpc",    HostPort: 0,    ContainerPort: 6450, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: nil,    Env: []v1.EnvVar{    {    Name: "POD_NAMESPACE",    Value: "",    ValueFrom: &v1.EnvVarSource{    FieldRef: &v1.ObjectFieldSelector{ -  APIVersion: "v1", +  APIVersion: "",    FieldPath: "metadata.namespace",    },    ResourceFieldRef: nil,    ConfigMapKeyRef: nil,    ... // 2 identical fields    },    },    {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}},    {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 5, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-55945d5dd", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-55945d5dd", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:13:03.394Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "22c19580-f8f8-4b87-a5d7-49676f745cdd", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "a16d388a-1d1c-4e2f-9fde-9f7a21700583", +  UID: "", -  ResourceVersion: "1777482709001183022", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiIwNzVkOTU0MzY1MGUwMjQ5Y2E4NzE5Yjc2OTYyYTM1NCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMDk3NjMyMjI0YmJmOWU5ZDQ1NGYyOWRmNDI4YTJiMzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiIwNzVkOTU0MzY1MGUwMjQ5Y2E4NzE5Yjc2OTYyYTM1NCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMDk3NjMyMjI0YmJmOWU5ZDQ1NGYyOWRmNDI4YTJiMzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:11:47 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:11:48 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-69b7cdbb99", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-69b7cdbb99", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:13:03.496Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "22c19580-f8f8-4b87-a5d7-49676f745cdd", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "a16d388a-1d1c-4e2f-9fde-9f7a21700583", +  UID: "", -  ResourceVersion: "1777482709001183022", +  ResourceVersion: "", -  Generation: 2, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiIwNzVkOTU0MzY1MGUwMjQ5Y2E4NzE5Yjc2OTYyYTM1NCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMDk3NjMyMjI0YmJmOWU5ZDQ1NGYyOWRmNDI4YTJiMzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiIwNzVkOTU0MzY1MGUwMjQ5Y2E4NzE5Yjc2OTYyYTM1NCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMDk3NjMyMjI0YmJmOWU5ZDQ1NGYyOWRmNDI4YTJiMzUifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:11:47 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:11:48 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{    Replicas: &0,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 5 identical fields    Ports: nil,    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: []v1.EnvVar{    ... // 3 identical elements    {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}},    {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}},    {    Name: "SCHEDULER_CHECKTIMEOUT", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_WRITERALSOREADER", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYUP", -  Value: "0", +  Value: "1",    ValueFrom: nil,    },    {    Name: "SCHEDULER_RETRYDOWN", -  Value: "0", +  Value: "3",    ValueFrom: nil,    },    {    Name: "SCHEDULER_PINGTIMEOUT", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_NODECHECKINTERVAL", -  Value: "0", +  Value: "2000",    ValueFrom: nil,    },    {    Name: "SCHEDULER_MAXCONNECTIONS", -  Value: "0", +  Value: "1000",    ValueFrom: nil,    },    {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"},    {Name: "PXC_READ_ONLY", Value: "false"},    },    Resources: {},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 2, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-69b7cdbb99", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-69b7cdbb99", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:13:08.169Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9890c72f-eee8-4ce0-a817-5e2e0d05ee53", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:08.169Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9890c72f-eee8-4ce0-a817-5e2e0d05ee53", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:08.170Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9890c72f-eee8-4ce0-a817-5e2e0d05ee53", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:08.170Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9890c72f-eee8-4ce0-a817-5e2e0d05ee53", "cluster": "pitr"} 2026-04-29T17:13:13.214Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8f1b6ee8-9106-4336-9ca1-0dfb4363c1d4", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:13.214Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8f1b6ee8-9106-4336-9ca1-0dfb4363c1d4", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:13.214Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8f1b6ee8-9106-4336-9ca1-0dfb4363c1d4", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:13.214Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "8f1b6ee8-9106-4336-9ca1-0dfb4363c1d4", "cluster": "pitr"} 2026-04-29T17:13:18.248Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3896a6-15aa-44ad-8ce2-ad6ac6722aef", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:18.248Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3896a6-15aa-44ad-8ce2-ad6ac6722aef", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:18.248Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3896a6-15aa-44ad-8ce2-ad6ac6722aef", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:18.248Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cb3896a6-15aa-44ad-8ce2-ad6ac6722aef", "cluster": "pitr"} 2026-04-29T17:13:23.267Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3d453c99-a793-4d67-9b47-78845903a5e5", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:23.267Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3d453c99-a793-4d67-9b47-78845903a5e5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:23.268Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3d453c99-a793-4d67-9b47-78845903a5e5", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:23.268Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3d453c99-a793-4d67-9b47-78845903a5e5", "cluster": "pitr"} 2026-04-29T17:13:28.292Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ed5cf2f6-3a61-479c-b2d7-9972b2bccbc9", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:28.292Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ed5cf2f6-3a61-479c-b2d7-9972b2bccbc9", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:28.293Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ed5cf2f6-3a61-479c-b2d7-9972b2bccbc9", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:28.293Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ed5cf2f6-3a61-479c-b2d7-9972b2bccbc9", "cluster": "pitr"} 2026-04-29T17:13:33.314Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7f6a6635-13fc-4684-b2f3-a4a88b8377b8", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:33.314Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7f6a6635-13fc-4684-b2f3-a4a88b8377b8", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:33.314Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7f6a6635-13fc-4684-b2f3-a4a88b8377b8", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:33.314Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7f6a6635-13fc-4684-b2f3-a4a88b8377b8", "cluster": "pitr"} 2026-04-29T17:13:38.332Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "afaee61a-0f31-4456-9fed-73e968f6d48c", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:38.332Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "afaee61a-0f31-4456-9fed-73e968f6d48c", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:38.333Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "afaee61a-0f31-4456-9fed-73e968f6d48c", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:38.333Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "afaee61a-0f31-4456-9fed-73e968f6d48c", "cluster": "pitr"} 2026-04-29T17:13:43.353Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "24263d04-6667-4302-9926-548dedf2367c", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:43.353Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "24263d04-6667-4302-9926-548dedf2367c", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:43.354Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "24263d04-6667-4302-9926-548dedf2367c", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:43.354Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "24263d04-6667-4302-9926-548dedf2367c", "cluster": "pitr"} 2026-04-29T17:13:45.679Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "aa45033a-6c41-4b14-8452-fcbd213c2286"} 2026-04-29T17:13:45.723Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "aa45033a-6c41-4b14-8452-fcbd213c2286", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:13:48.392Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b2783229-0a3a-44f7-8e78-966cf28487c3", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:48.392Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b2783229-0a3a-44f7-8e78-966cf28487c3", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:48.392Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b2783229-0a3a-44f7-8e78-966cf28487c3", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:48.392Z INFO point-in-time recovering {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b2783229-0a3a-44f7-8e78-966cf28487c3", "cluster": "pitr"} 2026-04-29T17:13:48.462Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "91ed237a-d6a9-4cf0-9d1a-122f51217060", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:13:50.896Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "83f58c96-6c2e-4684-b29f-ec3bd5cfd3df"} 2026-04-29T17:13:50.941Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "83f58c96-6c2e-4684-b29f-ec3bd5cfd3df", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:13:53.489Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d44d6433-d893-4503-8ed6-b643bd8b6bf7", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:13:56.125Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ce13f589-4567-4bf9-a8b9-fefb54fd4afe", "user": "xtrabackup"} 2026-04-29T17:13:56.132Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ce13f589-4567-4bf9-a8b9-fefb54fd4afe"} 2026-04-29T17:13:56.178Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "ce13f589-4567-4bf9-a8b9-fefb54fd4afe", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:13:58.509Z INFO starting cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3f6497c5-a872-468e-8473-a913890c75bd", "cluster": "pitr"} 2026-04-29T17:13:58.773Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "aaf476e9-b488-4ad4-8a80-b04851684ba9", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "19913d33-4030-4351-9d8e-6a7dd7c7a98b", +  UID: "", -  ResourceVersion: "1777482822466991024", +  ResourceVersion: "", -  Generation: 6, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:42 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &1, +  Replicas: &3,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", ...}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {Name: "tmp", VolumeSource: {EmptyDir: &{}}},    {    Name: "config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "auto-config",    VolumeSource: v1.VolumeSource{    ... // 16 identical fields    FC: nil,    AzureFile: nil,    ConfigMap: &v1.ConfigMapVolumeSource{    LocalObjectReference: {Name: "auto-pitr-pxc"},    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    VsphereVolume: nil,    Quobyte: nil,    ... // 9 identical fields    },    },    {    Name: "vault-keyring-secret",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "some-name-vault",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-users-secret-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "internal-pitr",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "mysql-init-file",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-mysql-init",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: nil,    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "grpc",    HostPort: 0,    ContainerPort: 6450, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: nil,    Env: []v1.EnvVar{    {    Name: "POD_NAMESPACE",    Value: "",    ValueFrom: &v1.EnvVarSource{    FieldRef: &v1.ObjectFieldSelector{ -  APIVersion: "v1", +  APIVersion: "",    FieldPath: "metadata.namespace",    },    ResourceFieldRef: nil,    ConfigMapKeyRef: nil,    ... // 2 identical fields    },    },    {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}},    {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"},    },    Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}},    ResizePolicy: nil,    ... // 6 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 3 identical fields    Args: {"mysqld"},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "sst",    HostPort: 0,    ContainerPort: 4444, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "write-set",    HostPort: 0,    ContainerPort: 4567, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "ist",    HostPort: 0,    ContainerPort: 4568, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysql-admin",    HostPort: 0,    ContainerPort: 33062, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "mysqlx",    HostPort: 0,    ContainerPort: 33060, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...},    ... // 4 identical fields    VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...},    VolumeDevices: nil,    LivenessProbe: &v1.Probe{    ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}},    InitialDelaySeconds: 300,    TimeoutSeconds: 5, -  PeriodSeconds: 10, +  PeriodSeconds: 0,    SuccessThreshold: 1,    FailureThreshold: 3,    TerminationGracePeriodSeconds: nil,    },    ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...},    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &600,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-pxc", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 6, +  ObservedGeneration: 0, -  Replicas: 1, +  Replicas: 0, -  ReadyReplicas: 1, +  ReadyReplicas: 0, -  CurrentReplicas: 1, +  CurrentReplicas: 0, -  UpdatedReplicas: 1, +  UpdatedReplicas: 0, -  CurrentRevision: "pitr-pxc-55945d5dd", +  CurrentRevision: "", -  UpdateRevision: "pitr-pxc-55945d5dd", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil, -  AvailableReplicas: 1, +  AvailableReplicas: 0,    },   } 2026-04-29T17:13:58.834Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "aaf476e9-b488-4ad4-8a80-b04851684ba9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "a16d388a-1d1c-4e2f-9fde-9f7a21700583", +  UID: "", -  ResourceVersion: "1777482783631647022", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-567565bcf7", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-567565bcf7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:13:58.905Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "aaf476e9-b488-4ad4-8a80-b04851684ba9", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true}   &v1.StatefulSet{    TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"},    ObjectMeta: v1.ObjectMeta{    ... // 2 identical fields    Namespace: "pitr-pxb-531",    SelfLink: "", -  UID: "a16d388a-1d1c-4e2f-9fde-9f7a21700583", +  UID: "", -  ResourceVersion: "1777482783631647022", +  ResourceVersion: "", -  Generation: 3, +  Generation: 0, -  CreationTimestamp: v1.Time{Time: s"2026-04-29 16:52:22 +0000 UTC"}, +  CreationTimestamp: v1.Time{},    DeletionTimestamp: nil,    DeletionGracePeriodSeconds: nil,    Labels: nil, -  Annotations: map[string]string{ -  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., -  }, +  Annotations: map[string]string{ +  "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., +  },    OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "c4ff554f-667c-4747-be2d-c2456a413847", ...}},    Finalizers: nil, -  ManagedFields: []v1.ManagedFieldsEntry{ -  { -  Manager: "kube-controller-manager", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., -  Subresource: "status", -  }, -  { -  Manager: "percona-xtradb-cluster-operator", -  Operation: "Update", -  APIVersion: "apps/v1", -  Time: s"2026-04-29 17:13:03 +0000 UTC", -  FieldsType: "FieldsV1", -  FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., -  }, -  }, +  ManagedFields: nil,    },    Spec: v1.StatefulSetSpec{ -  Replicas: &0, +  Replicas: &2,    Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Template: v1.PodTemplateSpec{    ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "075d9543650e0249ca8719b76962a354", "percona.com/ssl-internal-hash": "097632224bbf9e9d454f29df428a2b35"}},    Spec: v1.PodSpec{    Volumes: []v1.Volume{    {    Name: "ssl-internal",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl-internal",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &true,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {    Name: "ssl",    VolumeSource: v1.VolumeSource{    ... // 3 identical fields    AWSElasticBlockStore: nil,    GitRepo: nil,    Secret: &v1.SecretVolumeSource{    SecretName: "pitr-ssl",    Items: nil, -  DefaultMode: &420, +  DefaultMode: nil,    Optional: &false,    },    NFS: nil,    ISCSI: nil,    ... // 22 identical fields    },    },    {Name: "bin", VolumeSource: {EmptyDir: &{}}},    },    InitContainers: []v1.Container{    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    Containers: []v1.Container{    {    ... // 3 identical fields    Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...},    WorkingDir: "",    Ports: []v1.ContainerPort{    {    Name: "mysql",    HostPort: 0,    ContainerPort: 3306, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "proxyadm",    HostPort: 0,    ContainerPort: 6032, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    {    Name: "stats",    HostPort: 0,    ContainerPort: 6070, -  Protocol: "TCP", +  Protocol: "",    HostIP: "",    },    },    EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}},    Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...},    ... // 8 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    {    ... // 16 identical fields    StartupProbe: nil,    Lifecycle: nil, -  TerminationMessagePath: "/dev/termination-log", +  TerminationMessagePath: "", -  TerminationMessagePolicy: "File", +  TerminationMessagePolicy: "",    ImagePullPolicy: "Always",    SecurityContext: nil,    ... // 3 identical fields    },    },    EphemeralContainers: nil, -  RestartPolicy: "Always", +  RestartPolicy: "",    TerminationGracePeriodSeconds: &30,    ActiveDeadlineSeconds: nil, -  DNSPolicy: "ClusterFirst", +  DNSPolicy: "",    NodeSelector: nil,    ServiceAccountName: "default", -  DeprecatedServiceAccount: "default", +  DeprecatedServiceAccount: "",    AutomountServiceAccountToken: nil,    NodeName: "",    ... // 7 identical fields    Subdomain: "",    Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, -  SchedulerName: "default-scheduler", +  SchedulerName: "",    Tolerations: nil,    HostAliases: nil,    ... // 6 identical fields    PreemptionPolicy: nil,    Overhead: nil, -  TopologySpreadConstraints: nil, +  TopologySpreadConstraints: []v1.TopologySpreadConstraint{},    SetHostnameAsFQDN: nil,    OS: nil,    ... // 6 identical fields    },    },    VolumeClaimTemplates: []v1.PersistentVolumeClaim{    {    TypeMeta: {},    ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}},    Spec: v1.PersistentVolumeClaimSpec{    ... // 3 identical fields    VolumeName: "",    StorageClassName: nil, -  VolumeMode: &"Filesystem", +  VolumeMode: nil,    DataSource: nil,    DataSourceRef: nil,    VolumeAttributesClassName: nil,    },    Status: v1.PersistentVolumeClaimStatus{ -  Phase: "Pending", +  Phase: "",    AccessModes: nil,    Capacity: nil,    ... // 5 identical fields    },    },    },    ServiceName: "pitr-proxysql-unready", -  PodManagementPolicy: "OrderedReady", +  PodManagementPolicy: "",    UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, -  RevisionHistoryLimit: &10, +  RevisionHistoryLimit: nil,    MinReadySeconds: 0, -  PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", +  PersistentVolumeClaimRetentionPolicy: nil,    Ordinals: nil,    },    Status: v1.StatefulSetStatus{ -  ObservedGeneration: 3, +  ObservedGeneration: 0,    Replicas: 0,    ReadyReplicas: 0,    CurrentReplicas: 0,    UpdatedReplicas: 0, -  CurrentRevision: "pitr-proxysql-567565bcf7", +  CurrentRevision: "", -  UpdateRevision: "pitr-proxysql-567565bcf7", +  UpdateRevision: "", -  CollisionCount: &0, +  CollisionCount: nil,    Conditions: nil,    AvailableReplicas: 0,    },   } 2026-04-29T17:14:01.204Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "aaf476e9-b488-4ad4-8a80-b04851684ba9", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:14:03.612Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fedcab90-902d-404f-a512-f1a263635a85", "cluster": "pitr"} 2026-04-29T17:14:08.627Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "64cd6dfb-e873-47c1-83ec-992c8549a871", "cluster": "pitr"} 2026-04-29T17:14:09.480Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "9ee3a943-1fc5-4a74-ba7d-f3be1ac828f1", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-531 on 34.118.224.10:53: no such host"} 2026-04-29T17:14:13.685Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e8692ce7-009b-4497-bafa-9cca7a8ac2c6", "cluster": "pitr"} 2026-04-29T17:14:18.710Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ad0d15e9-bbf3-478e-a572-134ee35316e3", "cluster": "pitr"} 2026-04-29T17:14:23.727Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "d62194f7-2dca-449c-ae3b-d454e330c3e8", "cluster": "pitr"} 2026-04-29T17:14:28.871Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4bd7ce18-727e-475d-9dc5-11c09f3ad344", "cluster": "pitr"} 2026-04-29T17:14:33.890Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1c76b3fb-05e7-4469-aed3-21baa5ae4df9", "cluster": "pitr"} 2026-04-29T17:14:38.906Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5c153a91-e55b-45fb-90f6-a6189584915b", "cluster": "pitr"} 2026-04-29T17:14:43.928Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9001b1f7-0f6a-4d92-8922-088f7359a835", "cluster": "pitr"} 2026-04-29T17:14:48.948Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "484980d4-a81f-45c2-adaf-865c0f3b0ac7", "cluster": "pitr"} 2026-04-29T17:14:53.970Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "82301299-1550-4a55-97de-f4a2feca2ef9", "cluster": "pitr"} 2026-04-29T17:14:58.986Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1419e2ce-d71b-41fd-875a-def9ea008cc8", "cluster": "pitr"} 2026-04-29T17:15:04.005Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7fb75b7c-9c89-4ae0-804b-c7ab322717ac", "cluster": "pitr"} 2026-04-29T17:15:09.026Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "77e6af30-ad49-4f09-aa14-0ed9e9431dc5", "cluster": "pitr"} 2026-04-29T17:15:14.053Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1e934db6-9750-49fb-bb79-bb735855083f", "cluster": "pitr"} 2026-04-29T17:15:19.074Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "11178c80-4b21-46c1-8521-a42520715f70", "cluster": "pitr"} 2026-04-29T17:15:24.093Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c87cc9fd-be6a-4a0b-9860-60bab1bdfe6a", "cluster": "pitr"} 2026-04-29T17:15:29.133Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "adaeb894-84fc-4db4-a186-0980fa035991", "cluster": "pitr"} 2026-04-29T17:15:34.150Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2271c3db-fdd3-479e-a60d-7223ee6c593f", "cluster": "pitr"} 2026-04-29T17:15:39.169Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e1875012-9bea-4a44-94f1-f2b413ef57b2", "cluster": "pitr"} 2026-04-29T17:15:44.188Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3ab87fa5-7f5a-4b19-825c-c569118c2570", "cluster": "pitr"} 2026-04-29T17:15:49.207Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "da9fc15c-fcb6-437e-8ae9-b40fa84b1019", "cluster": "pitr"} 2026-04-29T17:15:54.226Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b9c16756-ef9b-4f6a-9dbc-ad6ead07c94e", "cluster": "pitr"} 2026-04-29T17:15:59.252Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "564a11e4-ace1-4379-bc90-20961559baec", "cluster": "pitr"} 2026-04-29T17:16:04.270Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "de875bcc-8b39-4391-a5b4-8a279eecb1de", "cluster": "pitr"} 2026-04-29T17:16:09.290Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "ff28dd0b-e1cf-4415-afad-0777ce9c989b", "cluster": "pitr"} 2026-04-29T17:16:14.308Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "fc1fffc5-7cd9-44bf-b681-476757a729ed", "cluster": "pitr"} 2026-04-29T17:16:19.326Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5007c168-156c-4df9-a67a-86e36de0a8b7", "cluster": "pitr"} 2026-04-29T17:16:24.353Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3589d52a-2079-441b-8433-064ba19bd3f2", "cluster": "pitr"} 2026-04-29T17:16:29.374Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "aff128d1-4511-4cb5-b3a4-186745eb0bde", "cluster": "pitr"} 2026-04-29T17:16:34.394Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2437a5fd-164f-4f7f-ad70-f99740bbfb46", "cluster": "pitr"} 2026-04-29T17:16:39.416Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "restore-on-pitr-minio-gtid", "reconcileID": "155fa02c-ddb9-4110-af00-95045010089b", "cluster": "pitr"} 2026-04-29T17:16:44.513Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "38cb2d32-7a47-4537-8244-4feb3e101c80"} 2026-04-29T17:16:48.271Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "edcdb5d7-c92f-4522-84b6-005d76e6aab2", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T17:16:50.379Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "edcdb5d7-c92f-4522-84b6-005d76e6aab2"} 2026-04-29T17:16:55.395Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "f71f4105-5bd0-427f-ae6f-0ffbda35fbdb"} 2026-04-29T17:16:59.998Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "9744cde6-34e5-4f15-b787-e5db4e646369", "latest": "2026-04-29 17:14:45 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:17:01.168Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "9744cde6-34e5-4f15-b787-e5db4e646369"} 2026-04-29T17:17:06.966Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "0d2684f8-f997-4d24-a7ec-626839207ff9"} 2026-04-29T17:17:12.593Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "5cda5cad-701b-41a8-ad1c-701d7990b14c"} 2026-04-29T17:17:18.689Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d41d62b9-59a6-4782-88f0-97b4c70a1f4a"} 2026-04-29T17:17:24.395Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "10b9de56-5aeb-4b01-bdb1-86caf0f8fbe0"} 2026-04-29T17:17:30.464Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "d1797e66-ba98-4088-9822-93c3e30f83f0"} 2026-04-29T17:17:36.105Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "f65d8457-5157-419c-aa0f-86c308a2f271"} 2026-04-29T17:17:41.778Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "7a48632e-f506-4f11-9668-b442cbb90b18"} 2026-04-29T17:17:47.686Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "27431555-9c09-4bed-aa9b-fee8998d37ff"} 2026-04-29T17:17:53.184Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "fcad5b87-62a6-4497-81c9-223f0313f7ec"} 2026-04-29T17:17:59.067Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "9179a87e-faeb-41c9-ac0a-5cc94845d4b6"} 2026-04-29T17:18:04.503Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "9bbc4ccd-e71b-45a6-8202-7fcdbe60ecc6"} 2026-04-29T17:18:10.510Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "61ddbf15-20cc-4fd7-a80a-1ba0874bf9bc"} 2026-04-29T17:18:16.303Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a8383559-a843-4afa-8fb6-8f3e787f8c12"} 2026-04-29T17:18:21.960Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a6a15c42-28eb-4c99-9279-e9d309826041"} 2026-04-29T17:18:27.470Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "22046fe0-a304-42dd-ae61-ec98ef55a31f"} 2026-04-29T17:18:33.579Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "52c555c4-44ed-41a7-b46b-b6ae1fdf25bd"} 2026-04-29T17:18:39.491Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "c090402a-d0f2-4884-b7a9-e5feb7191781"} 2026-04-29T17:18:44.910Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "87a77387-fb6f-4dc7-ad5c-6931b4a20aad"} 2026-04-29T17:18:51.302Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "034d24f9-e053-4c66-9f0d-f8a4cae93291"} 2026-04-29T17:18:57.086Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "9e661b52-9b5e-4651-bc8b-437056ea6917"} 2026-04-29T17:19:02.279Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "95cbacb6-279c-418d-b813-2dd4e8b3fc2c"} 2026-04-29T17:19:08.192Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "c0a8b928-1336-41ba-a96b-f36047d4bc4f"} 2026-04-29T17:19:13.589Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6cae1c28-13cc-432e-97c3-8bd1d65f4005"} 2026-04-29T17:19:19.285Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "e238c000-c596-47cb-806e-e8e89e6d4ae6"} 2026-04-29T17:19:25.175Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "75013dda-9f80-4801-bd7e-292607b34bf2"} 2026-04-29T17:19:30.969Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "30f75129-dd3f-4d43-bbd8-6653c801c4b9"} 2026-04-29T17:19:36.693Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "a9c43891-2d7b-44e0-85a7-132388422b22"} 2026-04-29T17:19:42.781Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "8a991924-593c-4474-a72b-8f17a74e7040"} 2026-04-29T17:19:48.902Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "674470a6-b59b-4f8e-a00d-bf7eaa97c47d"} 2026-04-29T17:19:54.477Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "6314e48a-376d-4045-993f-be41142e14b2"} 2026-04-29T17:20:00.094Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "4a9021b3-8266-4f07-82ed-7ea8046d2da3"} 2026-04-29T17:20:05.985Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "17d4f878-7a93-41ec-8de9-c049f739c4c0"} 2026-04-29T17:20:12.207Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "b8904ec5-4920-445b-b624-95bb290198a6"} 2026-04-29T17:20:17.987Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "5c489ad2-5838-4538-9c89-9f88e0ecb040"} 2026-04-29T17:20:23.419Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-531"}, "namespace": "pitr-pxb-531", "name": "pitr", "reconcileID": "44b76c97-0afb-428a-a76d-cd1a8b048563"} max retry count 120 reached. something went wrong with operator or kubernetes cluster